handoflixue comments on SotW: Be Specific - Less Wrong

37 Post author: Eliezer_Yudkowsky 03 April 2012 06:11AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (306)

Sort By: Controversial

You are viewing a single comment's thread. Show more comments above.

Comment author: handoflixue 05 April 2012 09:12:06PM 1 point [-]

the Genie will attempt to implement the wish in a way the results in a net decrease of utility for the wisher, but is bound by any constraints explicitly written into the wish

Constraint: This must result in a net increase in utility for me...

Comment author: Eliezer_Yudkowsky 06 April 2012 12:25:07AM 7 points [-]

I rewire your preferences. Oh, that wasn't what you meant by "utility"?

Comment author: Ezekiel 06 April 2012 12:48:11AM 8 points [-]

Incidentally, I find it funny (although not necessarily significant) that everyone else's instinct was to talk about the Genie in the third person, whereas Eliezer used the first person.

(Double-posted because it's a completely separate and much more frivolous comment.)

Comment author: shminux 08 April 2012 06:45:57PM 3 points [-]

It is extremely significant. That's partly the reason why EY managed to play the AI-in-a-box game rather successfully despite the overwhelming odds.

Comment author: [deleted] 08 April 2012 07:07:27PM 0 points [-]

Er.... How do you know? I thought he hadn't disclosed anything about how he did that.

Comment author: shminux 08 April 2012 08:24:41PM *  1 point [-]

He mentioned on some mailing list that he had to think like an AI desperately trying to get out. It makes a world of difference in how you approach the situation if it is your life that is actually on the line.

Comment author: TheOtherDave 08 April 2012 08:04:29PM 0 points [-]

The role of "the Genie" here and "the AI" in the Boxed AI game have certain obvious similarities.
It seems reasonable to assume that a willingness to adopt the former correlates with a willingness to adopt the latter.
It seems reasonable to assume that a willingness to adopt the role of "the AI" in the Boxed AI game is necessary (though not sufficient) in order to win that game.
So shminux's claim seems fairly uncontroversial to me.
Do you dispute it, or are you merely making a claim about the impossibility of knowledge?

Comment author: [deleted] 08 April 2012 08:18:48PM 0 points [-]

Do you dispute it, or are you merely making a claim about the impossibility of knowledge?

The latter.

Comment author: Ezekiel 06 April 2012 12:45:02AM 2 points [-]

You could try: "Constraint: The value that my current utility function would assign to the universe after this wish is implemented must be higher than the value my current utility function would assign to the universe that would have existed had you not implemented this wish."

... Which probably causes the Genie to, at best, throw an undefined-error, since human beings don't have well-defined utility functions. Since it's malicious, it will probably search through all of your desires, pick one of them at random to count as "my utility function", and then reinterpret the body of the wish to maximise that one thing at the expense of all others.

Comment author: wedrifid 06 April 2012 02:13:38AM *  2 points [-]

Since it's malicious, it will probably search through all of your desires, pick one of them at random to count as "my utility function", and then reinterpret the body of the wish to maximise that one thing at the expense of all others.

It's malicious and omnipotent. It'll do far worse than that. It'll scan your preferences until it finds a contradiction. Once you have a contradiction you can derive absolutely anything. It would then proceed to calculate your Coherent Extrapolated Volition and minimise it. It may not be obliged to figure out what you actually want but it can certainly do so for the purpose of being spiteful!

Comment author: Grognor 07 April 2012 07:17:35PM 6 points [-]

I think that is the first time I've ever seen anyone accurately describe the worst thing that could possibly happen.

Comment author: Dmytry 06 April 2012 01:04:48PM 1 point [-]

What happens if you ask it to maximize your CEV, though?

Lemme remember, the idea with CEV was what you'd desired if you thought faster and more reliably. Okay I ponder what will happen to you if your mind was BusyBeaver(10) times faster (way scarier number than 3^^^^3), without your body working any faster. 1 second passes.

Comment author: wedrifid 07 April 2012 02:14:20AM 2 points [-]

What happens if you ask it to maximize your CEV, though?

It'll fuck with you. Because that is what it does. It has plenty of scope to do so because CEV is not fully defined as of now. I'm not sure precisely how it would go about doing so. I just assume it does in some way I haven't thought of yet.

The meaning it attributes to CEV when it wants to exploit it to make things terrible is very different to the meaning it attributes to CEV when we try to use it to force it to understand us. It's almost as bad as some humans in that regard!

Comment author: Dmytry 07 April 2012 04:49:48AM 0 points [-]

It has plenty of scope to do so because CEV is not fully defined as of now.

The understatement of the year. CEV is vaguest crap ever with lowest hope of becoming less vague.

Comment author: wedrifid 07 April 2012 07:08:58AM 1 point [-]

with lowest hope of becoming less vague.

That's a rather significant claim.

Comment author: Dmytry 07 April 2012 07:16:18AM *  -1 points [-]

It's very uncommon to see crap this vague in development for such a long time by such a clever person, without it becoming less vague.

Comment author: wedrifid 07 April 2012 08:46:32AM *  1 point [-]

As far as I am aware this crap isn't in development. It isn't the highest research priority so the other SingInst researchers haven't been working on it much and Eliezer himself is mostly focused on writing a rationality book. Other things like decision theory are being worked on - which has involved replacing vague as crap TDT with less-vague UDT and UDT2.

I would like to see more work published on CEV. The most recent I am familiar with is this.

Comment author: TheOtherDave 06 April 2012 02:20:14AM 2 points [-]

Or alter my preferences so I antiprefer whatever it is able to produce the most of. Plus altering my brain such that my disutility and dishedonism are linear with that thing. Getting the attention of a Crapsack God sucks.

Comment author: wedrifid 06 April 2012 02:36:25AM 1 point [-]

Or alter my preferences so I antiprefer whatever it is able to produce the most of. Plus altering my brain such that my disutility and dishedonism are linear with that thing.

Those are actually subsumed under "mimimize CEV<TheOtherDave>". In the same way that maximising our CEV will not involve modifying our preferences drastically (unless it turns out we are into that sort of thing after all), minimising CEV would, if that turns out to be the worst way to @#$@ with us.

Getting the attention of a Crapsack God sucks.

Can't argue with that.