Psy-Kosh comments on The Sword of Good - Less Wrong

85 Post author: Eliezer_Yudkowsky 03 September 2009 12:53AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (292)

You are viewing a single comment's thread. Show more comments above.

Comment author: Psy-Kosh 03 September 2009 06:01:34PM 6 points [-]

Something that occurred to me along these lines. (not directly the same, but "close enough" that some of the moral judgments would be equivalent)

Let's say, next week, someone actually solved the mind uploading problem. They have a decision to make: go for it themselves, find someone as trustworthy as possible, forget about the plan and simply wait however long for the FAI math to be solved, etc...

What would you advise? Should they go for it themselves, try to then work out how to incrementally upgrade themselves without absolute disaster, forget it, etc etc etc...? (If nothing else, assume they already have the raw computing power to run a human at a vast speedup)

It's not an identical problem, but it's probably the closest thing.

Comment author: Eliezer_Yudkowsky 03 September 2009 07:22:11PM 12 points [-]

go for it themselves

What, you mean try to self-modify? Oh hell no. Human brain not designed for that. But you would have a longer time to try to solve FAI. You could maybe try a few non-self-modifications if you could find volunteers, but uploading and upload-driven-upgrading is fundamentally a race between how smart you get and how insane you get.

Comment author: MichaelVassar 05 September 2009 03:55:18AM 14 points [-]

The modified people can be quite a bit smarter than you are too, so long as you can see their minds and modify them. Groves et al managed to mostly control the Manhattan project despite dozens of its scientists being smarter than any of their supervisors and many having communist sympathies. If he actually shared their earlier memories and could look inside their heads... There's a limit to control, you still won't control an adversarial super intelligence this way, but a friendly human who appreciates your need for power over them? I bet they can have a >50 IQ point advantage, maybe even >70. Schoolteachers control children who have 70 IQ points on them with the help of institutions.

Comment author: Douglas_Knight 05 September 2009 06:03:37AM 8 points [-]

Schoolteachers control children who have 70 IQ points on them with the help of institutions.

Is it relevant that IQ is correlated with obedience to authority?

And how dumb do you think schoolteachers are? Bottom of those with BAs. I'd guess 100. And correlated with their pupils.

Comment author: Desrtopa 01 June 2011 06:29:00PM 3 points [-]

Estimations from SAT scores imply that the IQ of teachers and education majors is below average. Conscientious, hardworking students can graduate from most high schools and colleges with good grades, even if they are fairly stupid, as long as they stay away from courses which demand too much of them, and there are services available for those who are neither hardworking nor conscientious.

Education major courses are somewhat notorious for demanding little of students, and it is a stereotypically common choice for students seeking MRS degrees.

I'd like to imagine that the system would at least filter out individuals who are borderline retarded or below, but experience suggests to me that even this is too optimistic.

Comment author: Douglas_Knight 02 June 2011 03:01:53AM 3 points [-]

I don't buy the conversion in the first link, which is also a dead link. That Ed majors have an SAT score of 950 sounds right. That is 37th percentile among "college-bound seniors." If this population, which I assume means people taking the SAT, were representative of the general population, that would be an IQ of 95, but they aren't. I stand by my estimate of 100.

I doubt you have much experience with people with an IQ of 85, let alone the borderline retarded.

Comment author: Desrtopa 02 June 2011 06:42:55AM 3 points [-]

What makes you doubt I have much experience with either? IQ 85 is one standard deviation below average; close to 14 percent of the population has an IQ at least that low. The lower limit of borderline retardation, that is, the least intelligent you can be before you are no longer borderline, is two standard deviations below the mean, meaning that about one person in fifty is lower than that.

As it happens, I've spent a considerable amount of time with special needs students, some of whom suffer from learning disabilities which do not affect their reasoning abilities, but some of whom are significantly below borderline retarded.

At the public high school I attended, more than 95% of the students in my graduating year went on to college. While the most mentally challenged students in the area were not mainstreamed and didn't attend the same school, there was no shortage of <80 IQ students.

An average IQ of 100 for education majors would be within the error bars for the aforementioned projection, but some individuals are going to be considerably lower.

Comment author: Douglas_Knight 02 June 2011 07:27:23AM -2 points [-]

At the public high school I attended, more than 95% of the students in my graduating year went on to college. While the most mentally challenged students in the area were not mainstreamed and didn't attend the same school, there was no shortage of <80 IQ students.

Those two sentences are not very compatible.

Comment author: Desrtopa 02 June 2011 07:41:10AM 2 points [-]

The rates at which students progress to college have a lot more to do with parental expectations, funding, and the school environment than the intelligence of the students in question. My school had very good resources to support students in the admissions process, and students who didn't take it for granted that they were college bound were few and far between.

Comment author: Document 02 December 2010 11:55:54PM 1 point [-]

It seems unrealistic to assume that we'll be able to literally read the intentions of the first upload; I'd think that we'd start out not knowing any more about them than we would about an organic person through external scanning.

Comment author: MichaelVassar 03 December 2010 03:44:49AM 3 points [-]

You won't be able to evaluate their thoughts exactly, but there's a LOT that you should be able to tell about what a person is thinking if you can perfectly record all of their physiological reactions and every pattern of neural activation with perfect resolution, even with today's knowledge. Kock and Crick even found grandmother neurons, more or less.

Comment author: Document 03 December 2010 07:34:55AM 0 points [-]

I'd still expect it to be hard to tell the difference someone between thinking about or wanting to kill someone/take over the world and someone actually intending to. But I can imagine at least being able to reliably detect lies with that kind of information, so I'll defer to your knowledge of the subject.

Comment author: matt 06 September 2009 04:11:36AM *  13 points [-]

Eliezer, I'm with you that a properly designed mind will be great, but mere uploads will still be much more awesome than normal humans on fast forward.

Without hacking on how your mind fundamentally works, it seems pretty likely that being software would allow a better interface with other software than mouse, keyboard and display does now. Hacking on just the interface would (it seems to me) lead to improvements in mental capability beyond mere speed. This sounds like mind hacking to me (software enhancing a software mind will likely lead to blurry edges around which part we call "the mind"), and seems pretty safe.

Some (pretty safe*) cognitive enhancements:

  • Unmodified humans using larger displays are better at many tasks than humans using small displays (somewhat fluffy pdf research). It'll be pretty surprising if being software doesn't allow a better visual interface than a 30" screen.
  • Unmodified humans who can touch-type spend less time and attention on the mechanics of human machine interface and can be more productive (no research close to hand). Who thinks that uploaded humans are not going to be able to figure better interfaces than virtual keyboards?
  • Argument maps improve critical thinking, but the interfaces are currently clumsy enough to discourage use (lots of clicking and dragging). Who thinks that being software won't provide a better way to quickly generate argument maps?
  • In front of a computer loaded up with my keyboard shortcuts and browser plugins I have easy access to very fast lookup on various web reference sites. At the moment the lookup delay is still long enough that short term memory management (stack overflow after a mere 7±2 pushes) is a problem (when I need a reference I push my current task onto a mental stack; it takes time and attention to pop that task when the reference has been found). Who thinks I couldn't be smarter with a reference interface better than a keyboard?

All of which is just to say that I don't think you've tried very hard to think of safe self-modifications. I'm pretty confident that you could come up with more, and better, and safer than I have.

* Where "pretty safe" means "safe enough to propose to the LW community, but not safe enough to try before submitting for public ridicule"

Comment author: Psy-Kosh 03 September 2009 07:35:40PM *  3 points [-]

*blinks* I understand your "oh hell no" reaction to self modification and "use the speedup to buy extra time to solve FAI" suggestion.

However, I don't quite understand why you think "attempted upgrading of other" is all that much better. If you get that one wrong in a "result is super smart but insane (or, more precisely, very sane but with the goal architecture all screwed up) doesn't one end up with the same potential paths to disaster? At that point, if nothing else, what would stop the target from then going down the self modification path?

Comment author: Eliezer_Yudkowsky 03 September 2009 11:02:01PM 7 points [-]

Non-self-modification is by no means safe, but it's slightly less insanely dangerous than self-modification.

Comment author: Psy-Kosh 04 September 2009 12:35:20AM *  0 points [-]

Ooooh, okay then. That makes sense.

Hrm... given though your suggested scenario, why the need to start with looking for other volunteers? ie, if the initial person is willing to be modified under the relevant constraints, why not just, well, spawn off another instance of themselves, one the modifier and one the modifiee?

EDIT: whoops, just noticed that Vladimir suggested the same thing too.

Comment author: Nick_Tarleton 03 September 2009 08:10:40PM 6 points [-]

If insane happens before super-smart, you can stop upgrading the other.

Comment author: Psy-Kosh 03 September 2009 08:12:19PM 1 point [-]

Well, fair enough, there is that.

Comment author: pjeby 04 September 2009 04:16:15AM *  0 points [-]

What, you mean try to self-modify? Oh hell no. Human brain not designed for that

Perhaps you mean to say that we're not particularly trustworthy in our choices of what we modify ourselves to do or prefer?

Human brains, after all, are most exquisitely designed for modifying themselves, and can do it quite autonomously. They're just not very good at predicting the broader implications of those modifications, or at finding the right things to modify.

Comment author: Psy-Kosh 04 September 2009 06:26:35AM 2 points [-]

We're talking about direct explicit low level self modification. ie, uploading, then using that more convenient form to directly study one's own internal workings until one decides to go "hrm... I think I'll reroute these neural connections to... that, add a few more of this other kind of neuron over here and..."

Recall that the thing doing all that reasoning is the thing that's being affected by these modifications.

Comment author: pjeby 05 September 2009 01:10:04AM 3 points [-]

We're talking about direct explicit low level self modification. ie, uploading, then using that more convenient form to directly study one's own internal workings until one decides to go "hrm... I think I'll reroute these neural connections to... that, add a few more of this other kind of neuron over here and..."

Yes, but that would be the stupidest possible way of doing it, when there are already systems in place to do structured modification at a higher level of abstraction. Doing it at an individual neuron level would be like trying to... well, I would've said "write a property management program in Z-80 assembly," except I know a guy who actually did that. So, let's say, something about 1000 times harder. ;-)

What I find extremely irritating is when people talk about brain modification as if it's some sort of 1) terribly dangerous thing that 2) only happens post-uploading and 3) can only be done by direct hardware (or simulated hardware) modification. The correct answer is, "none of the above".

Comment author: CronoDAS 06 September 2009 03:17:12AM 4 points [-]

Well, we're talking about the kind of modifications that ordinary, non-invasive, high-level methods, acting through the usual sensory channels, don't allow. For example, no amount of ordinary self-help could make someone unable to feel physical pain, or can let you multiply large numbers extremely quickly in the manner of a savant. Changing someone's sexual orientation is also, at best, extremely difficult and at worst impossible. We can't seem to get rid of confirmation bias, or cure schizophrenia, or change an autistic brain into a neurotypical brain (or vice versa). There are lots of things that one might want to do to a brain that simply don't happen as long as that brain is sitting inside a skull only receiving input through normal human senses.

Comment author: Douglas_Knight 05 September 2009 06:05:40AM 2 points [-]

What I find extremely irritating is when people talk about brain modification as if it's some sort of 1) terribly dangerous thing that 2) only happens post-uploading and 3) can only be done by direct hardware (or simulated hardware) modification. The correct answer is, "none of the above".

Lists like that have a good chance of canceling out. That is, there are a bunch of ways people disagree with you because they're talking about something else.

Comment author: Vladimir_Nesov 03 September 2009 07:36:38PM *  1 point [-]

You can make volunteers out of your own copies. As long as the modified people aren't too smart, it's safe keep them in a sandbox and look through the theoretical work they produce on overdrive.

Comment author: matt 06 September 2009 04:17:17AM *  2 points [-]

AI boxes are pretty dangerous.

(I agree that "as long as the modified people aren't too smart" you're safe, but we are hacking on minds that will probably be able to hack on themselves, and possibly recursively self-improve if they decide, for instance, that they don't want to be shut down and deleted at the end of the experiment. I'm pretty strongly motiviated not to risk insanity by trying dangerous mind-hacking experiments, but I'm not going to be deleted in a few minutes.)

Comment author: cousin_it 03 September 2009 07:01:33PM *  1 point [-]

Difficult question. I believe those links are relevant, but your formulation also implies the threat of an arms race.

My best shot for now would be this: avoid self-modification. The top priority right now is defending people from the potential harmful effects of this thing you created, because someone less benevolent might stumble upon it soon. Find people who share this sentiment and use the speedup together to think hard about the problem of defense.

Comment author: Psy-Kosh 03 September 2009 07:28:40PM 0 points [-]

Perhaps an "anti arms race" would be a more accurate notion. ie, in once sense, waiting for the mathematics of FAI to be solved would be preferable. Would be safer to get to a point that we can mathematically ensure that the thing will be well behaved.

On the other hand, while waiting, how many will suffer and die irretrievably? If the cost for waiting was much smaller, then the answer of "wait for the math and construct the FAI rather than trying to patchwork update a spaghetti coded human mind" would be, to me, the clearly preferable choice.

Even given avoiding self modification, massive speedup would still correspond to significant amount of power. We already know how easily humans... change... with power. And when sped up, obviously people not sped up would seem different, "lesser"... helping to reinforce the "I am above them" sense. One might try to solve this by figuring out how to self modify enough to, well, not to that. But self modification itself being a starting point for, if one does not do it absolutely perfectly, potential disaster, well...

Anyways, so your suggestion would basically be "only use the power to, well, defend against the power" rather than use it to actually try to fix some of the annoying little problems in the world (like... death and and and and and... ?)

Comment author: cousin_it 03 September 2009 07:50:12PM *  2 points [-]

FAI is one possible means of defense, there might be others.

You shouldn't just wait for FAI, you should speed up FAI developers too because it's a race.

I think the strategy of developing a means of defense first has higher expected utility than fixing death first, because in the latter case someone else who develops uploading can destroy/enslave the world while you're busy fixing it.