MugaSofer comments on Rationality Quotes January 2013 - Less Wrong

6 Post author: katydee 02 January 2013 05:23PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (604)

You are viewing a single comment's thread. Show more comments above.

Comment author: Kawoomba 14 January 2013 06:16:13PM 1 point [-]

Since we're talking about CEV_(individual), the "poetic" definition would be "[my] wish if [I] knew more, thought faster, were more the [man] [I] wished [I] were, (...), where [my] wishes cohere rather than interfere; extrapolated as [I] wish that extrapolated, interpreted as [I] wish that interpreted."

Nothing that would change my top priorities, though I'd do a better job convincing Galactus.

Comment author: MugaSofer 14 January 2013 07:38:35PM -2 points [-]

How sure are you that you haven't made a mistake somewhere?

Comment author: Kawoomba 15 January 2013 05:55:34PM 5 points [-]

Quite sure. I assume you value the life of a sparrow (the bird), all else being equal. Is there a number of sparrows to spare which you would consign yourself and your loved ones to the flames? Is there a hypothetical number of sparrows for which you would choose them living over all currently living humans?

If not, then you are saying that not all goals reduce to a number on a single metric, that there are tiers of values, similar in principle to Maslow's.

You're my sparrows.

Comment author: JGWeissman 16 January 2013 02:49:30PM 7 points [-]

Suppose you had the chance to save the life of one sparrow, but doing so kills you with probability p. For what values of p would you do so?

If the answer is only when p=0, then your value of sparrows should never affect your choices, because it will always be dominated by the greater probability of your own welfare.

Comment author: Kawoomba 18 January 2013 12:07:13PM 2 points [-]

A strong argument, well done.

This indeed puts me in a conundrum: If I answer anything but p=0, I'm giving a kind of weighting factor that destroys the supposedly strict separation between tiers.

However, if I answer p=0, then indeed as long as there is anything even remotely or possibly affecting my top tier terminal values, I should rationally disregard pursuing any other unrelated goal whatsoever.

Obviously, as evident by my writing here, I do not solely focus all my life's efforts on my top tier values, even though I claim they outweigh any combination of other values.

So I am dealing with my value system in an irrational way. However, there are two possible conclusions concerning my confusion:

  • Are my supposed top tier terminal values in fact outweigh-able by others, with "just" a very large conversion coefficient?

or

  • Do I in fact rank my terminal values as claimed and am just making bad choices effectively matching my behavior to those values, wasting time on things not strictly related to my top values? (Is it just an instrumental rationality failure?) Anything with a terminal value that's valued infinitely higher than all other values should behave strictly isomorphically to a paperclip maximizer with just that one terminal value, at least in our universe.

This could be resolved by Omega offering me a straight out choice, pressing buttons or something. I know what my consciously reflected decision would be, even if my daily routine does not reflect that.

Another case of "do as I say (I'd do in hypothetical scenarios), not as I do (in daily life)" ...

Comment author: wedrifid 18 January 2013 02:57:28PM 0 points [-]

This indeed puts me in a conundrum: If I answer anything but p=0, I'm giving a kind of weighting factor that destroys the supposedly strict separation between tiers.

Well, you could always play with some fun math...

Comment author: [deleted] 18 January 2013 03:58:41PM 0 points [-]

Even that would be equivalent to an expected utility maximizer using just real numbers, except that there's a well-defined tie-breaker to be used when two different possible decisions would have the exact same expected utility.

Comment author: MugaSofer 20 January 2013 01:34:12PM *  -2 points [-]

How often do two options have precisely the same expected utility? Not often, I'm guessing. Especially in the real world.

Comment author: [deleted] 20 January 2013 05:57:47PM *  1 point [-]

I guess almost never (in the mathematical sense). OTOH, in the real world the difference is often so tiny that it's hard to tell its sign -- but then, the thing to do is gather more information or flip a coin.

Comment author: ArisKatsaris 16 January 2013 03:04:59PM -1 points [-]

If the answer is only when p=0, then your value of sparrows should never affect your choices, because it will always be dominated by the greater probability of your own welfare.

Not sure that holds. Surely there could be situations where you can't meaningfully calculate whether acting to preserve the life of a sparrow increases or decreases the probability of your death, therefore you act to preserve its life because though you consider it a fundamentally lesser terminal value, it's still a terminal value.

Comment author: JGWeissman 16 January 2013 03:18:40PM 0 points [-]

Surely there could be situations where you can't meaningfully calculate whether acting to preserve the life of a sparrow increases or decreases the probability of your death

In this case you try harder to figure out a way to calculate the impact on your chance of death. The value of information of such an effort is worth infinite sparrow lives. Lower tier utility functions just don't matter.

Comment author: ArisKatsaris 16 January 2013 03:39:49PM *  -1 points [-]

In this case you try harder to figure out a way to calculate the impact on your chance of death. The value of information of such an effort is worth infinite sparrow lives.

What if you've already estimated that calculating excessively (e.g. beyond a minute) on this matter will have near-definite negative impact on your well-being?

Comment author: JGWeissman 16 January 2013 03:52:07PM *  0 points [-]

Then you go do something else that's relevant to your top-tier utility function.

You can contrive a situation where the lower tier matters, but it looks like someone holding a gun to your head, and threatening to kill you if you don't choose in the next 5 seconds whether or not they shoot the sparrow. That sort of thing generally doesn't happen.

And even then, if you have the ability to self-modify, the costs of maintaining a physical representation of the lower tier utility functions is greater than the marginal benefit of choosing to save the sparrow automatically because you lower tier utility function says so over choosing alphabetically.

Comment author: BerryPick6 15 January 2013 06:08:18PM 2 points [-]

I think this is more due to diminishing marginal returns for the amount of sparrows in existence, to be honest...

Jokes aside, you are offering a very persuasive argument; I'd be curious to know how you figure out what tier certain values are and whether you ever have reason to (a) change your mind about said tier or (b) create new tiers altogether?

Comment author: MugaSofer 16 January 2013 02:40:32PM *  -2 points [-]

Interesting analogy. If we accept that utilities are additive, then there is presumably a number of sparrows worth killing for. (Of course, there may be a limit on all possible sparrows or sparrow utilities may be largely due to species preservation or something. As an ethics-based vegetarian, however, I can simply change it to "sparrows tortured.) I would be uncomfortable trying to put a number on it, what with the various sacred value conflicts involved, but I accept that a Friendly AI (even one Friendly only to me) would know and act on it.

Maslow's Pyramid is not intended as some sort of alternative to utilitarianism, it's a description of how we should prioritize the needs of humans. An imperfect one, of course, but better than nothing.

How sure? Based on what? What would persuade you otherwise?

Comment author: Eugine_Nier 17 January 2013 01:15:22AM 1 point [-]

If we accept that utilities are additive

Why?

Comment author: [deleted] 17 January 2013 01:51:22AM 1 point [-]

They almost certainly are on the margin (think taylor series approx of utility function). Get to the point where you are talking about killing a significant fraction of the sparrow population, then there's no reason to think so.

Comment author: Eugine_Nier 18 January 2013 04:34:19AM 2 points [-]

They almost certainly are on the margin

True, but this doesn't apply to MugaSofer's claim that

there is presumably a number of sparrows worth killing for.

Comment author: DaFranker 15 January 2013 06:35:39PM *  0 points [-]

If not, then you are saying that not all goals reduce to a number on a single metric, (...)

Simply false AFAIK. There is a mathematical way to express as a single-number metric every tiered system I've ever been capable of conceiving, and I suspect those I haven't also have such expressions with more mathematics I might not know.

So, I don't know if the grandparent was saying that, but I assume it wasn't, and if it was implied somewhere and I missed it then it certainly is false.

But then again, I may simply be interpreting your words uncharitably. I assume you're already aware that Maslow's can also be reduced to a single number formula.

A more interesting question than maximizing numbers of sparrows, however, is maximizing other value-factors. Suppose instead of imagining a number of sparrows sufficiently large for which you would trade a human you care about, you imagine a number of different minimum-sufficiency-level values being traded off for that "higher-tiered" life.

One human against a flock of sparrows large enough to guarantee the survival of the species is easy enough. Even the sparrow species doesn't measure up against a human, or at least that's what I expect you'd answer.

Now measure it up against a flock of each species of birds, on pain of extinction of birds (but let's magically handwave away the ecosystem impacts of this - suppose we have advanced technology to compensate for all possible effects).

Now against a flock of each species of birds, a group of each species of nonhuman mammals, a school of each species of fish, a colony of each species of insects, and so forth throughout the entire fauna and flora of the planet - or of all known species of life in the universe. Again we handwave - we can make food using the power of Science and so on.

Now against the same, but without the handwaving. The humans you care about are all fine and healthy, but the world they live in is less interesting, and quite devastated, but Science keeps y'all healthy and stuff.

Now against that, plus having actual human bodies. You're all jarbrains.

Feel free to keep going, value by value, until you reach a sufficient tradeoff or accept that no amount of destructive alteration to the universe will ever compare to the permanent loss of consciousness of your loved ones, and therefore you'd literally do everything and anything, up to and including warping space and time and sacrificing knowledge or memory or thinking-power or various qualia or experience or capacity for happiness or whatever else you can imagine, all traded for this one absolute value.

"Life/consciousness of loved ones" can also be substituted for whichever is your highest-tiered value if different.

Comment author: Kawoomba 15 January 2013 06:47:00PM 0 points [-]

Simply false AFAIK. There is a mathematical way to express as a single-number metric every tiered system I've ever been capable of conceiving

I don't understand. Do you mean that as in "you can describe/encode arbitrary systems as a single number" or something related to that?

If not, do you mean that there must be some number of sparrows outweighing everything else as it gets sufficiently large?

Please explain.

Comment author: DaFranker 15 January 2013 07:20:28PM *  0 points [-]

Do you mean that as in "you can describe/encode arbitrary systems as a single number" or something related to that?

Yes.

For my part, I also consider it perfectly plausible (though perhaps less likely than some alternatives) that some humans might actually have tiered systems where certain values really truly never can be traded off in the slightest fraction of opportunity costs against arbitrarily high values of all lower-tiered values at the same time.

For instance, I could imagine an agent that values everything I value but has a hard tier cutoff below the single value that its consciousness must remain continuously aware until the end of the universe if such a time ever arrives (forever otherwise, assuming the simplest alternative). This agent would have no trouble sacrificing the entire solar system if it was proven to raise the expected odds of this survival. Or the agent could also only have to satisfy a soft threshold or some balancing formula where a certain probability of eternal life is desired, but more certainty than that becomes utility-comparable to lower-tier values. Or many other kinds of possible constructs.

So yes, arbitrary systems, for all systems I've ever thought of. I like to think of myself as imaginative and as having thought of a lot of possible arbitrary systems, too, though obviously my search space is limited by my intelligence and by the complexity I can formulate.

Comment author: Kawoomba 15 January 2013 07:34:40PM 0 points [-]

There are actual tiered systems all around us, even if most examples that come to mind are constructed/thought of by humans.

That aside, I am claiming that I would not trade my highest tier values against arbitrary combinations of all lower tiered values. So ... hi!

Re: Just a number; I can encode your previous comments (all of them) in the form of a bitstring, which is a number. Doesn't mean that doing "+1" on that yields any sensible result. Maybe we're talking past each other on the "describe/encode" point, but I don't see how describing a system containing strict tiers as a number somehow makes those tiers go away, unless you were nitpicking about "everything's just a number that's interpreted in a certain way" or somesuch.

Comment author: DaFranker 15 January 2013 07:56:16PM *  0 points [-]

Ah, on the numbers thing, what I meant was only that AFAIK there always exists some formula for which higher output numbers will correspond to things any abitrary agent (at least, all the logically valid and sound ones that I've thought of) would prefer.

So even for a hard tier system, there's a way to compute a number linearly representative of how happy the agent is with worldstates, where at the extreme all lower-tier values flatline into arbitrarily large negatives (or other, more creative / leakproof weighing) whenever they incur infinitesimal risk of opportunity cost towards the higher-tier values.

The reason I'm said this is because it's often disputed and/or my audience isn't aware of it, and I often have to prove even the most basic versions of this claim (such as "you can represent a tiered system where as soon as the higher tier is empty, the lower tier is worthless using a relatively simple mathematical formula") by showing them the actual equations and explaining how it works.