Vladimir_Nesov comments on Open Thread June 2010, Part 3 - Less Wrong

6 Post author: Kevin 14 June 2010 06:14AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (606)

You are viewing a single comment's thread. Show more comments above.

Comment author: JoshuaZ 14 June 2010 10:02:09PM *  3 points [-]

I'm not aware of anyone here who would claim that LW is one of the most important things in the world right now but I think a lot of people here would agree that improving human reasoning is important if we can have those improvements apply to lots of different people across many different fields.

There is a definite group of people here who think that SIAI is really important. If one thinks that a near Singularity is a likely event then this attitude makes some sense. It makes a lot of sense if you assign a high probability to a Singularity in the near future and also assign a high probability to the possibility that many Singularitarians either have no idea what they are doing or are dangerously wrong. I agree with you that the SIAI is not that important. In particular, I think that a Singularity is not a likely event for the foreseeable future, although I agree with the general consensus here that a large fraction of Singularity proponents are extremely wrong at multiple levels.

Keep in mind that for any organization or goal, the people you hear the most about it are the people who think that it is important. That's the same reason that a lot of the general public thinks that tokamak fusion reactors will be practical in the next fifty years: The physicists and engineers who think that are going to loudly push for funding. The ones who don't are going to generally just go and do something else. Thus, in any given setting it can be difficult to estimate the general communal attitude towards something since the strongest views will be the views that are most apparent.

Comment author: Vladimir_Nesov 14 June 2010 10:24:49PM *  13 points [-]

I don't think intelligence explosion is imminent either. But I believe it's certain to eventually happen, absent the end of civilization before that. And I believe that its outcome depends exclusively on the values of the agents driving it, hence we need to be ready, with good understanding of preference theory at hand when the time comes. To get there, we need to start somewhere. And right now, almost nobody is doing anything in that direction, and there is very poor level of awareness of the problem and poor intellectual standards of discussing the problem where surface awareness is present.

Either right now, or 50, or 100 years from now, a serious effort has to be taken on, but the later it starts, the greater the risk of being too late to guide the transition in a preferable direction. The problem itself, as a mathematical and philosophical challenge, sounds like something that could easily take at least 100 years to reach clear understanding, and that is the deadline we should worry about, starting 10 years too late to finish in time 100 years from now.

Comment author: Benquo 15 June 2010 01:04:15AM -1 points [-]

"But I believe it's certain to eventually happen, absent the end of civilization before that."

And I will live 1000 years, provided I don't die first.

Comment author: Vladimir_Nesov 15 June 2010 01:26:46AM *  2 points [-]

But I believe it's certain to eventually happen, absent the end of civilization before that.

And I will live 1000 years, provided I don't die first.

(As opposed to gradual progress, of course. I could make a case with your analogy facing an unexpected distinction also, as in what happens if you got overrun by a Friendly intelligence explosion, and persons don't prove to be a valuable pattern, but death doesn't adequately describe the transition either, as value doesn't get lost.)

Comment author: multifoliaterose 15 June 2010 12:10:43AM 0 points [-]

Vladimir, I agree with you that people should be thinking intelligence explosion, that there's a very poor level of awareness of the problem, and that the intellectual standards for discourse about this problem in the general public are poor.

I have not been convinced but am open toward the idea that a paperclip maximizer is the overwhelmingly likely outcome if we create a superhuman AI. At present, my thinking is that if some care is taking in the creation of a superhuman AI, more likely than a paperclip maximizer is an AI which partially shares human values, that is, the dicotomy "paper clip maximizer vs. Friendly AI" seems like a false dicotomy - I imagine that the sort of AI that people would actually build would be somewhere in the middle. Any recommended reading on this point appreciated.

SIAI seems to have focused on the existential risk of "unfriendly intelligence explosion" and it's not clear to me that this existential risk is greater than the risks coming from world war and natural resource shortage.

Comment author: Vladimir_Nesov 15 June 2010 01:02:29AM 6 points [-]

the dichotomy "paper clip maximizer vs. Friendly AI" seems like a false dichotomy - I imagine that the sort of AI that people would actually build would be somewhere in the middle. Any recommended reading on this point appreciated.

Mainly Complexity of value. There is no way for human values to magically jump inside the AI, so if it's not specifically created to reflect them, it won't have them, and whatever the AI ends up with won't come close to human values, because human values are too complex to be resembled by any given structure that happens to be formed in the AI.

The more AI's preference diverges from ours, the more we lose, and this loss is on astronomic scale (even if preference diverges relatively little). The falloff with imperfect reflection of values might be so sharp that any ad-hoc solution turns the future worthless. Or maybe not, with certain classes of values that contain a component of sympathy that reflects values perfectly while giving them smaller weight in the overall game, but then we'd want to technically understand this "sympathy" to have any confidence in the outcome.

Comment author: CarlShulman 16 June 2010 04:33:09PM 4 points [-]

The more AI's preference diverges from ours, the more we lose, and this loss is on astronomic scale (even if preference diverges relatively little).

This depends on something like aggregative utilitarianism. If additional resources have diminishing marginal value in fulfilling human aims, that getting a little slice of the universe (in the course of negotiating terms of surrender with the inhuman AI, if it can make credible commitments, or because we serve as acausal bargaining chips with other civilizations elsewhere in the universe) may be enough. Is getting 100% of the lightcone a hundred times better than 1%?

Comment author: Wei_Dai 21 June 2010 01:27:33AM 9 points [-]

Is getting 100% of the lightcone a hundred times better than 1%?

I think yes, if we take into account that the more of the lightcone we (our FAI) get, the more trading opportunities we would have with UFAI in other possible worlds. Diminishing marginal value shouldn't apply across possible worlds, because otherwise it would imply gross violations of expected utility maximization.

Also, I suspect that there are possible worlds with much greater resources than our universe (perhaps with physics that allow hypercomputation, or just many orders of magnitude more total exploitable resources), and some of them would have potential trading partners who are willing to give us a small share of their world for a large share of ours. We may eventually achieve most of our value from trading with them. But of course such trade wouldn't be possible if we didn't have something to trade with!

Comment author: Vladimir_Nesov 21 June 2010 10:55:00AM *  6 points [-]

Interesting. This suggests thinking about FAI not as using its control to produce terminal value in its own world, but as using its control to buy as much terminal value as it can, in various world-programs. Since it doesn't matter where the value is produced, most of the value doesn't have to be produced in the possible worlds with FAIs in them. Indeed, it sounds unlikely that specifically the FAI worlds will be optimal for FAI-value optimization. FAIs (and the worlds they control) act as instrumental leverage, a way of controlling the global mathematical universe into having more value for our preference.

Thus, more FAIs means stronger control over the mathematical universe, while more UFAIs mean that the mathematical universe is richer, and so the FAIs can get more value out of it with the same control. The metaphors of trade and comparative advantage start applying again, not on the naive level of cohabitation on the same world, but on the level of the global ontology. Mathematics grants you total control over your domain, so that your "atoms" can't be reused for something else by another stronger agent, and so you do benefit from most superintelligent "aliens".

Comment author: Wei_Dai 21 June 2010 11:29:07AM *  5 points [-]

Yes, assuming that trading across possible worlds can be done in the first place. One thing that concerns me is the combinatorial explosion of potential trading partners. How do they manage to "find" each other?

Comment author: Vladimir_Nesov 21 June 2010 11:59:21AM *  4 points [-]

It's the same combinatorial explosion as with the future possible worlds. Even though you can't locate individual valuable future outcomes (through certain instrumental sequences of exact events), you can still make decisions about your actions leading to certain consequences "in bulk", and I expect the trade between possible worlds can be described similarly (after all, it does work on exactly the same decision-making algorithm). Thus, you usually won't know who are you trading with, exactly, but on the net estimate that your actions are in the right direction.

Comment author: JoshuaZ 17 May 2012 07:14:39PM 1 point [-]

Isn't the set of future worlds with high measure a lot smaller?

Comment author: Will_Newsome 27 September 2010 06:29:23AM 4 points [-]

Robin Hanson would be so pleased that it turns out economics is the fundamental law of the entire ensemble universe.

Comment author: Vladimir_Nesov 16 June 2010 04:51:03PM 1 point [-]

There are two distinct issues here: (1) how high would a human with original preference value a universe which only gives a small weight to their preference, and (2) how likely is the changed preference to give any weight whatsoever to the original preference, in other words to produce a universe to any extent valuable to the original preference, even if original preference values universes only weakly optimized in its direction.

Moving to a different preference is different from lowering weight of the original preference. A slightly changed (formal definition of) preference may put no weight at all on the preceding preference. The optimal outcome according to the modified preference can thus be essentially moral noise, paperclips, to the original preference. Giving a small slice of the universe, on the other hand, is what you get out of aggregation of preference, and a changed preference doesn't necessarily have a form of aggregation that includes original preference. (On the other hand, there is a hope that human-like preferences include sympathy, which does make them aggregate preferences of other persons with some weight.)

Comment author: CarlShulman 16 June 2010 04:56:26PM 1 point [-]

We should assign some substantial probability to getting some weighting of our preference (from bargaining with transparency, acausal trade, altruistic brain emulations, etc). If a moderate weighting of our preferences gets most of the potential utility, then the expected utility of inhuman AIs getting powerful won't be astronomically less than the expected utility of, e.g. a 'Friendly AI' acting on idealized human preferences.

Comment author: Vladimir_Nesov 16 June 2010 05:28:50PM 1 point [-]

We should assign some substantial probability to getting some weighting of our preference (from bargaining with transparency, acausal trade, altruistic brain emulations, etc).

Game-theoretic considerations are only relevant when you have non-trivial control, not when your atoms are used for something else. If singleton's preference gives some weight to your preference, this is a case of having control directly through the singleton's preference, but the origin of this control is not game-theoretic. If the singleton's preference has sympathy for your preference, your explicit instantiation in the world doesn't need to have any control, in order to win through the implicit control via singleton's preference.

Game-theoretic aggregation, on the other hand, doesn't work by influence on other agent's preference. You only get your slice of the world because you already control it. Another agent may perform trade, but this is trade of control, rearranging what specifically each of you controls, without changing your preferences.

I assume that control will be winner-takes-all, so preferences of other agents existing at the time only matter if the winner's preference directly pays to their preferences any attention, but not if they had some limited control from the start.

If a moderate weighting of our preferences gets most of the potential utility, then the expected utility of inhuman AIs getting powerful won't be astronomically less than the expected utility of, e.g. a 'Friendly AI' acting on idealized human preferences.

My point is that inhuman AI may give no weight to our preference, while FAI may give at least some weight to everyone's preference. Game-theoretic trade won't matter here because agents other than the singleton have no control to bargain with. FAI gives weight to other preferences not because of trade, but by construction from the start, even if people it gives weight to don't exist at all (FAI giving them weight in optimization might cause them to appear, or a better event at least as good from their perspective).

Comment author: CarlShulman 17 June 2010 06:21:47AM 3 points [-]

You only get your slice of the world because you already control it.

This isn't obviously the most natural way to describe a scenario in which an AI thinks it has a 90% chance of winning a conflict with humanity, but also has the ability to jointly create (with humanity) agents to enforce an agreement (and can do this quickly enough to be relevant), so cuts a deal splitting up the resources of the light cone at a 9:1 ratio.

I assume that control will be winner-takes-all,

Given that there are plausible sets of parameter values where this assumption is false, we can't use it to assess overall expected value to astronomical precision.

Game-theoretic considerations are only relevant when you have non-trivial control,

I specifically mentioned acausal trade, a la Rolf Nelson's AI-deterrence scheme, which needs non-trivial control only in some region of the ensemble of possibilities the AI considers. Indeed, the AI might treat us well simply because of the chance that benevolent non-human aliens will respond positively if its algorithm has this output (as the benevolent aliens might be modeling the AI's algorithm).

Comment author: Vladimir_Nesov 18 June 2010 04:11:23PM *  4 points [-]

I specifically mentioned acausal trade, a la Rolf Nelson's AI-deterrence scheme, which needs non-trivial control only in some region of the ensemble of possibilities the AI considers.

Yes, I forgot about that (though I remain uncertain about how well this argument works, not having worked out a formal model). To summarize the arguments for why future is still significantly more valuable than what we have now, even if we run into Unfriendly AI,

(1) if there is a non-negligible chance that we'll have FAI in the future, or that we could've created FAI if some morally random facts in the past (such as the coin in counterfactual mugging) were different, then we can estimate the present expected value of the world as pretty high, as a factor of getting whole universes (counterfactually or probably) optimized towards your specific preference is present in the expected utility computation. The counterfactual value is present even if it's certain that the future contains Unfriendly AI.

(2) It's even better, because the unfriendly singletons will also optimize their worlds towards your preference a little for game-theoretic reasons, even if they don't care at all about your preference. This game is not with you personally, a human that controls very little and whose control can't compel a singleton to any significant extent, but with the counterfactual FAIs. The FAIs that could be created, but weren't, can act as Omega in counterfactual mugging, making it profitable for the indifferent singletons to pay the FAI a little in FAI-favored kind of world-optimization.

(3) Some singletons that don't follow your preference in particular, but have remotely human-like preference, will have a component of sympathy in their preference, and will dole your preference some fair portion of control in their world, that is much greater than the portion of control you held originally. This sympathy seems to be godshatter of game-theoretic considerations that compel even singletons with non-evolved (artificial, random) preferences according to arguments (1) and (2).

The conclusion to this seems to be that creating an Unfriendly AI is significantly better than ending up with no rational singleton at all (existential disaster that terminates civilization), but significantly worse than a small chance of FAI.

Comment author: multifoliaterose 15 June 2010 08:36:09AM *  1 point [-]

There is no way for human values to magically jump inside the AI, so if it's not specifically created to reflect them, it won't have them, and whatever the AI ends up with won't come close to human values, because human values are too complex to be resembled by any given structure that happens to be formed in the AI.

I'm not convinced by the claim that human values have high Kolmogorov complexity.

In particular, Eliezer's article Not for the Sake of Happiness Alone is totally at odds with my own beliefs. In my mind, it's incoherent to give anything other than subjective experiences ethical consideration. My own preference for real science over imagined science is entirely instrumental and not at all terminal.

Now, maybe Eliezer is confused about what his terminal values are, or maybe I'm confused about what my terminal values are, or maybe our terminal values are incompatible. In any case, it's not obvious that an AI should care about anything other than the subjective experiences of sentient beings.

Suppose that it's okay for an AI to exclude everything but subjective experience from ethical consideration. Is there then still reason to expect that human values have high Kolmogorov complexity?

I don't have a low complexity description to offer, but it seems to me that one can get a lot of mileage out of the principles "if an individual prefers state A to state B whenever he/she/it is in either of state A or state B, then state A is superior for that individual to state B" and "when faced with two alternatives, the moral alternative is the one that you would prefer if you were going to live through the lives of all sentient beings involved."

Of course "sentient being" is ill-defined and one would have to do a fair amount of work frame the things that I just said in more formal terms, but anyway, it's not clear to me that there's a really serious problem here.

The more AI's preference diverges from ours, the more we lose, and this loss is on astronomic scale (even if preference diverges relatively little).

I totally agree that if the creation of a superhuman AI is going to precede all other existential threats then we should focus all of our resources on trying to get the superhuman AI to be as friendly as possible.

Comment author: khafra 15 June 2010 10:28:27AM 3 points [-]

Have you read the Heaven post by denisbider and the two follow-ups constituting a mini-wireheading series? There have been other posts on the difference between wanting and liking; but it illustrates a fairly strong problem with wireheading: Even if all we're worried about is "subjective states," many people won't want to be put in that subjective state, even knowing they'll like it. Forcing them into it or changing their value system so they do want it are ethically suboptimal solutions.

So, it seems to me that if anything other than maximized absolute wireheading for everyone is the AI's goal, it's gonna start to get complicated.

Comment author: multifoliaterose 15 June 2010 11:26:16PM 0 points [-]

Thanks for the references to the posts which I had not seen before and which I find relevant. I'm sympathetic toward denisbider's view, but will read the comments to see if I find diverging views compelling.

Comment author: Vladimir_Nesov 15 June 2010 11:16:42AM 1 point [-]

Maybe you should start with what's linked from fake fake utility functions then (the page on the wiki wasn't organized quite as I expected).

Comment author: Vladimir_Nesov 15 June 2010 11:08:40AM 0 points [-]

I totally agree that if the creation of a superhuman AI is going to precede all other existential threats then we should focus all of our resources on trying to get the superhuman AI to be as friendly as possible.

(?) I never said that.

Comment author: multifoliaterose 15 June 2010 07:32:19PM 0 points [-]

Yes, I was agreeing with what I inferred your attitude to be rather than agreeing with something that you said. (I apologize if I distorted your views - if you'd like I can edit my comment to remove the suggestion that you hold the position that I attributed to you.)

Comment author: Vladimir_Nesov 15 June 2010 08:05:25PM 3 points [-]

I don't believe that we "should focus all of our resources" on FAI, as there are many other worthy activities to focus on. The argument is that this particular problem gets disproportionally little attention, and while with other risks we can in principle luck out even if they get no attention, it isn't so for AI. Failing to take FAI seriously is fatal, failing to take nanotech seriously isn't necessarily fatal.

Thus, although strictly speaking I agree with your implication, I don't see its condition plausible, and so implication as whole relevant.

Comment author: timtyler 16 June 2010 03:35:48PM *  0 points [-]

Re: "Is there then still reason to expect that human values have high Kolmogorov complexity?"

Human values are mosly a product of their genes and their memes. There is an awful lot of information in those. However, it is true that you can fairly closely approximate human values - or those of any other creature - by the directive to make as many grandchildren as possible - which seems reasonably simple.

Most of the arguments for humans having complex values appear to list a whole bunch of proximate goals - as though that constitutes evidence.

Comment author: SilasBarta 16 June 2010 04:34:53PM *  3 points [-]

I disagree. You need to know much more than just the drive for grandchildren, given the massively diverse ways we observe even in our present world for species to propagate, all of which correspond to different articulable values once they reach human intelligence.

Human values should be expected to have a high K-complexity because you would need to specify both the genes/early environment, and the precise place in history/Everett branches where humans are now.

Comment author: timtyler 16 June 2010 04:43:11PM *  0 points [-]

The idea was to "approximate human values" - not to express them in precise detail: nobody cares much if Jim likes strawberry jam more than he likes raspberry jam.

The environment mostly drops out of the equation - because most of it is shared between the agents involved - and because of the phenomenon of Canalisation: http://en.wikipedia.org/wiki/Canalisation_%28genetics%29

Comment author: SilasBarta 16 June 2010 04:49:01PM *  1 point [-]

The idea was to "approximate human values" - not to express them in precise detail

Sure, but I take "approximation" to mean something like getting you within 10 or so bits of the true distribution, but the heuristic you gave still leaves you maybe 500 or so bits away, which is huge, and far more than you implied.

The environment mostly drops out of the equation - because most of it is shared between the agents involved - and because of the phenomenon of Canalisation

That would help you on message length if you had already stored one person's values and were looking to store a second person's. It does not for describing the first person's value, or some aggregate measure of humans' values.

Comment author: timtyler 16 June 2010 04:55:50PM 1 point [-]

10 bits!!! That's not much of a message!

The idea of a shared environment arises because the proposed machine - in which the human-like values are to be implemented - is to live in the same world as the human. So, one does not need to specify all the details of the environment - since these are shared naturally between the agents in question.

Comment author: multifoliaterose 15 June 2010 08:41:44AM *  0 points [-]

But I would qualify the last sentence of my reply by saying that the best way to get a superhuman AI to be as friendly as possible may not be to work on friendly AI or advocate for friendly AI. For example, it may be best to work toward geopolitical stability to minimize the chances of some country rashly creating a potentially unsafe AI out of a sense of desperation during wartime.

Comment author: Vladimir_Nesov 15 June 2010 01:58:31AM *  5 points [-]

SIAI seems to have focused on the existential risk of "unfriendly intelligence explosion" and it's not clear to me that this existential risk is greater than the risks coming from world war and natural resource shortage.

Not clear to me either that unfriendly AI is the greatest risk, in the sense of having the most probability of terminating the future (though "resource shortage" as existential risk sounds highly implausible - we are talking about extinction risks, not merely potential serious issues; and "world war" doesn't seem like something particularly relevant for the coming risks, dangerous technology doesn't need war to be deployed).

But Unfriendly AI seems to be the only unavoidable risk, something we'd need to tackle in any case if we get through the rest. On other problems we can luck out, not on this one. Without solving this problem, the efforts to solve the rest are for naught (relatively speaking).

Comment author: multifoliaterose 15 June 2010 04:44:00AM 3 points [-]

"resource shortage" as existential risk sounds highly implausible - we are talking about extinction risks, not merely potential serious issues;

I mean "existential risk" in a broad sense.

Suppose we run out of a source of, oh, say, electricity too fast to find a substitute. Then we would be forced to revert to a preindustrial society. This would be a permanent obstruction to technological progress - we would have no chance of creating a transhuman paradise or populating the galaxy with happy sentient machines and this would be an astronomical waste.

Similarly if we ran out of any number of things (say, one of the materials that's currently needed to build computers) before finding an adequate substitute.

"world war" doesn't seem like something particularly relevant for the coming risks, dangerous technology doesn't need war to be deployed.

My understanding is that a large scale nuclear war could seriously damage infrastructure. I could imagine this preventing technological development as well.

But Unfriendly AI seems to be the only unavoidable risk, something we'd need to tackle in any case if we get through the rest. On other problems we can luck out, not on this one. Without solving this problem, the efforts to solve the rest are for naught (relatively speaking).

On the other hand, it's equally true that if another existential risk hits us before we friendly AI, all of our friendly AI directed efforts will be for naught.

Comment author: Strange7 15 June 2010 08:52:42AM 3 points [-]

Suppose we run out of a source of, oh, say, electricity too fast to find a substitute.

That's not how economics works. If one source of electricity becomes scarce, that means it's more expensive, so people will switch to cheaper alternatives. All the energy we use ultimately comes from either decaying isotopes (fission, geothermal) or the sun; neither of those will run out in the next thousand years.

Modern computer chips are doped silicon semiconductors. We're not going to run out of sand any time soon, either. Of course, purification is the hard part, but people have been thinking up clever ways to purify stuff since before they stopped calling it 'alchemistry.'

Comment author: khafra 15 June 2010 01:37:45PM 3 points [-]

The energy requirements for running modern civilization aren't just a scalar number--we need large amounts of highly concentrated energy, and an infrastructure for distributing it cheaply. The normal economics of substitution don't work for energy.

A "tradeoff" exists between using resources (including energy and material inputs of fossil origin) to feed the growth of material production (industry and agriculture) and to support the economy’s structural transformation.

As the substitution of renewable for nonrenewable (primarily fossil) energy continues, nature exerts resistance at some point; the scale limit begins to bind. Either economic growth or transition must halt. Both alternatives lead to severe disequilibrium. The first because increased pauperization and the apparent irreducibility of income differentials would endanger social peace. Also, since an economic order built on competition among private firms cannot exist without expansion, the free enterprise system would flounder.

The second alternative is equally untenable because the depletion of nonrenewable resources, proceeding along a rising marginal cost curve or, equivalently, along a descending Energy Return on Energy Invested (EROI) schedule, increases production costs across the entire spectrum of activities. Supply curves shift upwards.

It's entirely possible that failure to create a superintelligence before the average EROI drops too low for sustainment would render us unable to create one for long enough to render other existential risks inevitabilities.

Comment author: timtyler 16 June 2010 08:34:00AM *  0 points [-]

"Substitution economics" seems unlikely to stop us eventually substituting fusion power and biodesiel for oil. Meanwhile, we have an abundance of energy in the form of coal - more than enough to drive progress for a loooog while yet. The "energy apocalypse"-gets-us-first scenario is just very silly.

Comment author: khafra 16 June 2010 01:32:43PM *  2 points [-]

Energy economics is interconnected enough with politics to make me lower my expectation of rationality from both of us for the remainder of the discussion due to reference class forecasting. Also, we are several inferential steps away from each other, so any discussion is going to be long and full of details. Regardless, I'm going to go ahead, assuming agreement that market forces cannot necessarily overcome resource shortages (or the Easter Islanders would still be with us).

Historically, the world switched from coal to petroleum before developing any technologies we'd regard as modern. The reason, unlike so much else in economics, is simple: the energy density of coal is 24 MJ/kg; the energy density of gasoline is 44 MJ/kg. Nearly doubling the energy density makes many things practical that wouldn't otherwise be, like cars, trucks, airplanes, etc. Coal cannot be converted into a higher energy density fuel except at high expense and with large losses, making the expected reserves much smaller. The fuels it can be converted to require significant modifications to engines and fuel storage.

Coal is at least plausible, although a stop-gap measure with many drawbacks. It's your hopes for fusion that really show the wishful thinking. Fusion is 20 years away from being a practical energy source, just like it was in 1960. The NIF has yet to reach break-even; economically practical power generation is far beyond that point; assuming a substantial portion of US energy generation needs is farther still. It'd be nice if Polywell/Bussard fusion proved practical, but that's barely a speck on the horizon, getting its first big basic research grant from the US Navy. And nothing but Mr. Fusion will help unless someone makes an order of magnitude improvement in battery or ultracapacitor energy density.

No matter which of the alternatives you plan to replace the energy infrastructure with, you needed to start about 20 years ago. World petroleum production is no longer sufficient to sustain economic growth and infrastructure transition simultaneously. Remember, the question isn't whether it's theoretically possible to substitute more plentiful energy sources for the ones that are getting more difficult to extract, it's whether the declining EROI of current energy sources will remain high enough for the additional economic activity of converting infrastructure to other sources while still feeding people, let alone indulging in activities with no immediate payoff like GAI research.

We seem to be living in a world where the EROI is declining faster than willingness to devote painful amounts of the GDP to energy source conversion is increasing. This doesn't mean an immediate biker zombie outlaw apocalypse, but it does mean a slow, unevenly distributed "catabolic collapse" of decreasing standards of living, security, and stability.

Comment author: RobinZ 18 June 2010 08:07:20PM 1 point [-]

Upvoted chiefly for

Energy economics is interconnected enough with politics to make me lower my expectation of rationality from both of us for the remainder of the discussion due to reference class forecasting.

but I appreciate the analysis. (I am behind on reading comments, so I will be continuing downthread now.)

Comment author: JoshuaZ 16 June 2010 05:26:01PM 0 points [-]

And nothing but Mr. Fusion will help unless someone makes an order of magnitude improvement in battery or ultracapacitor energy density.

I don't know why you focus so much on fusion although I agree it isn't practical at this point. But note that batteries and ultracapacitors are just energy storage devices. Even if they become far more energy dense they don't provide a source of energy.

Comment author: timtyler 16 June 2010 05:05:02PM *  -1 points [-]

That we will eventually substitute fusion power and biodesiel for oil seems pretty obvious to me. You are saying it represents "wishful thinking" - because of the possibility of civilisation not "making it" at all? If so, be aware that I think that the chances of that happening seem to grossly exaggerated around these parts.

Comment author: cupholder 15 June 2010 11:06:08AM 2 points [-]

That's not how economics works. If one source of electricity becomes scarce, that means it's more expensive, so people will switch to cheaper alternatives.

I would have thought that those 'cheaper alternatives' could still be more expensive than the initial cost of the original source of electricity...? In which case losing that original source of electricity could still bite pretty hard (albeit maybe not to the extent of being an existential risk).

Comment author: Vladimir_Nesov 15 June 2010 11:02:17AM 1 point [-]

On the other hand, it's equally true that if another existential risk hits us before we friendly AI, all of our friendly AI directed efforts will be for naught.

Yes.

Comment author: CarlShulman 16 June 2010 04:27:29PM 1 point [-]

But Unfriendly AI seems to be the only unavoidable risk, something we'd need to tackle in any case if we get through the rest.

A stably benevolent stable world government/singleton could take its time solving AI, or inching up to it with biological and culture intelligence enhancement. From our perspective we should count that as almost a maximal win in terms of existential risks.

Comment author: Vladimir_Nesov 16 June 2010 05:00:05PM 3 points [-]

I don't see your point. It would take an unrealistic world dictatorship (whether it's "benevolent" seems like irrelevant hair-splitting at that point) to stop the risks (stop the technological progress in the wild!) and allow more time for development of FAI. And in the end, solving FAI still remains a necessary step, even if done by modified/improved people, even if given a safe environment to work in.

Comment author: CarlShulman 16 June 2010 05:24:52PM 1 point [-]

I don't see your point. It would take an unrealistic world dictatorship (whether it's "benevolent" seems like irrelevant hair-splitting at that point) to stop the risks (stop the technological progress in the wild!) and allow more time for development of FAI.

You were talking about hundred year time scales. That's time enough for neuroscience lie detectors, whole brain emulation, democratization in authoritarian countries, continued expansion of EU-like arrangements, and many other things to occur. That's time for lie detectors/neuroscience to advance a lot, whole brain emulation to take off

And in the end, solving FAI still remains a necessary step, even if done by modified/improved people, even if given a safe environment to work in.

But from our perspective, if we can get the benevolent non-AI (but perhaps WBE) singleton, it can do the FAI work at leisure and we don't need to. So the relative marginal impact of our working on say, FAI theory or institutional arrangements for WBE, need to be weighed against one another.

Comment author: Vladimir_Nesov 16 June 2010 05:59:05PM 2 points [-]

You were talking about hundred year time scales. That's time enough for neuroscience lie detectors, whole brain emulation, democratization in authoritarian countries, continued expansion of EU-like arrangements, and many other things to occur. That's time for lie detectors/neuroscience to advance a lot, whole brain emulation to take off

It's also time enough for any of the huge number of other outcomes. It's not outright impossible, but pretty improbable, that the world will go this exact road. And don't underestimate how crazy people are.

But from our perspective, if we can get the benevolent non-AI (but perhaps WBE) singleton, it can do the FAI work at leisure and we don't need to. So the relative marginal impact of our working on say, FAI theory or institutional arrangements for WBE, need to be weighed against one another.

After the change of mind about value of drifted human preference, I agree that WBE/intelligence enhancement is a viable road. Here're my arguments about the impact of these paths at this point.

WBE is still at least decades away, probably more than a hundred years if you take planning fallacy into account, and depends on the development of global technological efforts that are not easily influenced. Value of any "institutional arrangements" and viability of arguing for them given the remoteness (hence irrelevance at present) and implausibility (to most people) of WBE, also seems doubtful at present. This in my mind makes the marginal value on any present effort related to WBE relatively small. This will go up sharply as WBE tech gets closer

I suspect that FAI theory, once understood, will still be simple enough (if any general theory is possible), and can be developed by vanilla humans (on unknown timescale, probably decades to hundreds of years, but at some point WBEs overtake the timescale estimates). By the time WBE becomes viable, the risk situation will be already very explosive, so if we can get a good understanding earlier, we could possibly avoid that risky period entirely. Also, having a viable technical Friendliness programme might give academic recognition to the problem (that these risks are as unavoidable as laws of physics, and not just something to talk with your friends about, like politics or football), which might spread awareness of the AI risks on an otherwise unachievable level, helping with institutional change promoting measures against wild AI and other existential risks. On the other hand, I won't underestimate human craziness on this point as well - technical recognition of the problem may still live side to side with global indifference.

Comment author: Craig_Morgan 15 June 2010 03:35:31AM *  4 points [-]

I have not been convinced but am open toward the idea that a paperclip maximizer is the overwhelmingly likely outcome if we create a superhuman AI. At present, my thinking is that if some care is taking in the creation of a superhuman AI, more likely than a paperclip maximizer is an AI which partially shares human values, that is, the dicotomy "paper clip maximizer vs. Friendly AI" seems like a false dicotomy - I imagine that the sort of AI that people would actually build would be somewhere in the middle. Any recommended reading on this point appreciated.

I believed similarly until I read Steve Omohundro's The Basic AI Drives. It convinced me that a paperclip maximizer is the overwhelmingly likely outcome of creating an AGI.

Comment author: CarlShulman 16 June 2010 04:24:26PM *  2 points [-]

That paper makes a convincing case that the 'generic' AI (some distribution of AI motivations weighted by our likelihood of developing them) will most prefer outcomes that rank low in our preference ordering, i.e. the free energy and atoms needed to support life as we know it or would want it will get reallocated to something else. That means that an AI given arbitrary power (e.g. because of a very hard takeoff, or easy bargaining among AIs but not humans, or other reasons) would be lethal. However, the situation seems different and more sensitive to initial conditions when we consider AIs with limited power that must trade off chances of conquest with a risk of failure and retaliation. I'm working on a write up of those issues.

Comment author: multifoliaterose 15 June 2010 06:35:54AM 1 point [-]

Thanks Craig, I'll check it out!