XiXiDu comments on Evidence for the orthogonality thesis - Less Wrong

11 Post author: Stuart_Armstrong 03 April 2012 10:58AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (289)

You are viewing a single comment's thread. Show more comments above.

Comment author: Will_Newsome 03 April 2012 11:00:01PM *  8 points [-]

A lot of the arguments given in these comments amount to: We can imagine a narrow AI that somehow becomes a general intelligence without wireheading or goal distortion, or, We can imagine a specific AGI architecture that is amenable to having precisely defined goals, and because we can imagine them, they're probably possible, and if they're probably possible, then they're probable. But such an argument is very weak. Our intuitions might be wrong, those AIs might not be the first to be developed, they might be theoretically possible but not pragmatically possible, and so on. Remember, we still don't know what intelligence is! We can define it as cross-domain optimization or what have you, but such definitions are not automatically valid just because they look sorta math-y. AIXI is probably not intelligent in the sense that a human is intelligent, and thus won't be dangerous. Why should I believe that any other AI architectures you come up with on the fly are any more dangerous?

So whenever you say, "imagine an AIXI approximation with a specific non-friendly utility function: that would be bad!", my response is, "who says such an AGI is even possible, let alone probable?". And whenever you say, "Omohundro says...", my response is, "Omohundro's arguments are informal and suggestive, but simply nowhere near conclusive, and in fact parts of his arguments can be taken to suggest in favor of an AI detecting and following moral law". There just aren't any knock-down arguments, because we don't know what it takes to make AGI. The best you can do is to make pragmatic arguments that caution is a good idea because the stakes are high. When people in this community act as if they have knock-down arguments where there aren't any it makes SingInst and LessWrong look like weirdly overconfident end-of-the-world-mongers.

(Also, the 'AGI will literally kill us all by default' argument is laughably bad, for many game theoretic and economic reasons both standard and acausal that should be obvious, and people unthinkingly repeating it also makes SingInst and LessWrong look like weirdly overconfident end-of-the-world-mongers.)

Comment author: XiXiDu 04 April 2012 08:49:18AM *  2 points [-]

...because we can imagine them, they're probably possible, and if they're probably possible, then they're probable. [...] And whenever you say, "Omohundro says...", my response is, "Omohundro's arguments are informal and suggestive, but simply nowhere near conclusive...

Completely agree with your comment. Conceivability does not imply conceptuality, does not imply logical possibility, does not imply physical possibility, does not imply economic feasibility. Yet the arguments uttered on Less Wrong seldom go beyond conceivability.

When people in this community act as if they have knock-down arguments where there aren't any it makes SingInst and LessWrong look like weirdly overconfident end-of-the-world-mongers.

This is exactly the impression I got when I first started asking about risks from AI. Most of all comments I got have been incredible poor and without any substance. But the commentators do not notice that themselves because other people on lesswrong seemingly agree and they get upvoted. Yet nobody with the slightest doubts would be convinced.

All they manage to do is convince those who already hold the same set of beliefs or who fit a certain mindset.

The best you can do is to make pragmatic arguments that caution is a good idea because the stakes are high.

I just reread this post yesterday and found it to be a very convincing counter-argument against the idea that we should solely act on high stakes.

Comment author: Will_Newsome 04 April 2012 10:03:52AM 6 points [-]

All they manage to do is convince those who already hold the same set of beliefs or who fit a certain mindset.

It's perhaps worth noting that this observation is true of most discussion about most even-mildly-controversial subjects on LessWrong—quantum mechanics, cryonics, heuristics and biases, ethics, meta-ethics, theology, epistemology, group selection, hard takeoff, Friendliness, et cetera. What confuses me is that LessWrong continues to attract really impressive people anyway; it seems to be the internet's biggest/best forum for interesting technical discussion about epistemology, Schellingian game theory, the singularity, &c., even though most of the discussion is just annoying echoes. One of a hundred or so regular commenters is actually trying or is a real intellectual, not a fountain of cultish sloganeering and cheering. Others are weird hybrids of cheerleader and actually trying / real intellectual (like me, though I try to cheer on a higher level, and about more important things). Unfortunately I don't know of any way to raise the "sanity waterline", if such a concept makes sense, and I suspect that the new Center for Modern Rationality is going to make things worse, not better. I hope I'm wrong. ...I feel like there's something that could be done, but I have no idea what it is.

I just reread this post yesterday and found it to be a very convincing counter-argument against the idea that we should solely act on high stakes.

Eh, I think Vassar's reply is more to the point.

Comment author: Eugine_Nier 05 April 2012 01:33:36AM 1 point [-]

What confuses me is that LessWrong continues to attract really impressive people anyway; it seems to be the internet's biggest/best forum for interesting technical discussion about epistemology, Schellingian game theory, the singularity, &c., even though most of the discussion is just annoying echoes.

Why is that confusing? Have you looked at the rest of the internet recently?

Comment author: Will_Newsome 05 April 2012 03:34:16AM 4 points [-]

Have you looked at the rest of the internet recently?

Not really. But are you saying that nowhere else on the internet is close to LessWrong's standards of discourse? I'd figured that but part of me keeps saying "there's no way that can be true" for some reason.

I'm not sure why I'm confused, but I think there's a place where my model (of how many cool people there are and how willing they would be to participate on a site like LessWrong) is off by an order of magnitude or so.

Comment author: Eugine_Nier 05 April 2012 06:40:29AM 2 points [-]

how many cool people there are and how willing they would be to participate on a site like LessWrong

A better question is how many of them are willing to create a site like LessWrong.

Also minor nitpick about your use of the word 'cool', since it normally denotes social status rather than rationality.

Comment author: XiXiDu 05 April 2012 09:26:23AM 3 points [-]

Have you looked at the rest of the internet recently?

Not really. But are you saying that nowhere else on the internet is close to LessWrong's standards of discourse? I'd figured that but part of me keeps saying "there's no way that can be true" for some reason.

It might be true when it comes to cross-domain rationality (with a few outliers like social abilities). But it certainly isn't true that Less Wrong is anywhere close to the edge in most fields (with a few outliers like decision theory).

Comment author: XiXiDu 10 April 2012 02:58:56PM *  1 point [-]

I just reread this post yesterday and found it to be a very convincing counter-argument against the idea that we should solely act on high stakes.

Eh, I think Vassar's reply is more to the point.

I think Wei_Dai's reply does trump that.

What Vassar is saying sounds to me like a justification of Pascal's Wager by arguing that some God's have more measure than others and that therefore we can rationally decide to believe into a certain God and live accordingly.

That is like saying that a biased coin does not have a probability of 1/2 and that we can therefore maximize our payoff by betting on the side of the coin that is more likely to end up face-up. Which would be true if we had any other information other than that the coin is biased. But if we don't have any reliable information except other than that it is biased, it makes no sense to deviate from the probability of a fair coin.

And I don't think it is clear, at this point, that we are justified to assume more than that there might be risks from AI. Claiming that there are actions that we can take, with respect to risks from AI, that are superior to others, is like claiming that the coin is biased while being unable to determine the direction of the bias. By claiming that doing something is better than doing nothing we might as well end up making things worse. Just like by unconditionally assigning a higher probability to one side of a coin, of which we know nothing but that it is biased, in a coin tossing tournament.

The only sensible option seems to be to wait for more information.

Comment author: Rain 10 April 2012 08:17:11PM 1 point [-]

This is one of The Big Three Problems I came to LW hoping to find a solution for, but have mainly noticed that nobody wants to talk about it. Oh well.

Comment author: XiXiDu 11 April 2012 07:38:17PM 1 point [-]

This is one of The Big Three Problems I came to LW hoping to find a solution for, but have mainly noticed that nobody wants to talk about it. Oh well.

Now I am curious about the other two.

Comment author: Rain 12 April 2012 02:29:30AM 2 points [-]
  • How do you judge what you should (value-judgmentally) value?
  • How do you deal with uncertainty about the future (unpredictable chains of causality)? (what your above post was about)
  • What's the right thing to do in life?

Here are some of my previous posts on the topics.

Comment author: XiXiDu 12 April 2012 02:54:09PM *  1 point [-]

Here are some of my previous posts on the topics.

Your posts highlight fundamental problems that I have as well. Especially this and this comment concisely describe the issues.

I have no answers and I don't know how other people deal with it. Personally I forget about those problems frequently and act as if I can actually calculate what to do. Other times I just do what I want based on naive introspection.

Comment author: timtyler 13 April 2012 01:26:27PM *  0 points [-]

And I don't think it is clear, at this point, that we are justified to assume more than that there might be risks from AI. Claiming that there are actions that we can take, with respect to risks from AI, that are superior to others, is like claiming that the coin is biased while being unable to determine the direction of the bias. By claiming that doing something is better than doing nothing we might as well end up making things worse. Just like by unconditionally assigning a higher probability to one side of a coin, of which we know nothing but that it is biased, in a coin tossing tournament.

This is a problem - though it probably shouldn't stop us from trying.

The only sensible option seems to be to wait for more information.

Players can try to improve their positions and attempt to gain knowledge and power. That itself might cause problems - but it seems likely to beat thumb twiddling.

Comment author: lukstafi 04 April 2012 09:54:36PM 1 point [-]

Why do you think that "Center for Modern Rationality" is going to make things worse? Let's hope it will not hinge on Eliezer Yudkowsky's more controversial deliberations (as for me, his thoughts on: the complexity of ethical value, the nature of personhood, the solution to FAI).

Comment author: Will_Newsome 04 April 2012 10:12:48PM *  5 points [-]

I don't think what they teach will be particularly harmful to people's epistemic habits, but I don't think it'll be helpful either, and I think that there will be large selection effects for people who will, through sheer osmosis and association with the existent rationalist community, decide that it is "rational" to donate a lot of money to the Singularity Institute or work on decision theory. It seems that the Center for Modern Rationality aims to create a whole bunch of people at roughly the average LessWrong commenter level of prudence. LessWrong is pretty good relatively speaking, but I don't think their standards are nearly high enough to tackle serious problems in moral philosophy and so on that it might be necessary to solve in order to have any good basis for one's actions. I am disturbed by the prospect of an increasingly large cadre of people who are very gung-ho about "getting things done" despite not having a deep understanding of why those things might or might not be good things to do.