All of PrawnOfFate's Comments + Replies

I don't get it: any agent that fooms becomes superintelligent. It's values don't necessarily change at all, nor does its connection to its society.

2Randaly
An agent in a society is unable to force its values on the society; it needs to cooperate with the rest of society. A singleton is able to force its values on the rest of society.
4TheOtherDave
Sure. But any answer to that metaethical question which allows us to class some bases for comparison as moral values and others as merely values implicitly privileges a moral reference frame (or, rather, a set of such frames). Where I was going is that you asked me a question here which I didn't understand clearly enough to be confident that my answer to it would share key assumptions with the question you meant to ask. So I asked for clarification of your question. Given your clarification, and using your terms the way I think you're using them, I would say that whether it's valid to class a moral change as moral progress is a metaethical question, and whatever answer one gives implicitly privileges a moral reference frame (or, rather, a set of such frames). If you meant to ask me about my preferred metaethics, that's a more complicated question, but broadly speaking in this context I would say that I'm comfortable calling any way of preferentially sorting world-states with certain motivational characteristics a moral frame, but acknowledge that some moral frames are simply not available to minds like mine. So, for example, is it moral progress to transition from a social norm that in-practice-encourages randomly killing fellow group members to a social norm that in-practice-discourages it? Yes, not only because I happen to adopt a moral frame in which randomly killing fellow group members is bad, but also because I happen to have a kind of mind that is predisposed to adopt such frames.

UFAI is about singletons. If you have an AI society whose members compare notes and share information -- which ins isntrumentally useful for them anyway -- your reduce the probability of singleton fooming.

-2MugaSofer
Any agent that fooms becomes a singleton. Thus, it doesn't matter if they acted nice while in a society; all that matters is whether they act nice as a singleton.
0TheOtherDave
Asserting that some bases for comparison are "moral values" and others are merely "values" implicitly privileges a moral reference frame. I still don't understand what you mean when you ask whether it's valid to do so, though. Again: if I decide that this hammer is better than that hammer because it's blue, is that valid in the sense you mean it? How could I tell?

And is that valid or not? If you can validly decide some systems are better than others, you are some of the way to deciding which is best.

0TheOtherDave
Can you say more about what "valid" means here? Just to make things crisper, let's move to a more concrete case for a moment... if I decide that this hammer is better than that hammer because it's blue, is that valid in the sense you mean it? How could I tell?
-3MugaSofer
FAI is about singletons, because the first one to foom wins, is the idea. ETA: also, rational agents may be ethical in societies, but there's no advantage to being an ethical singleton.
-4MugaSofer
In my experience, most people don't think moral progress involves changing reference frames, for precisely this reason. If they think about it at all, that is.
5ArisKatsaris
That's hardly the best example you could have picked since there are obvious metrics by which South Africa can be quantifiably called a worse society now -- e.g. crime statistics. South Africa has been called the "crime capital of the world" and the "rape capital of the world" only after the fall of the Apartheid. That makes the lack of moral progress in South Africa a very easy bullet to bite - I'd use something like Nazi Germany vs modern Germany as an example instead.

I'm increasingly baffled as to why AI is always brought in to discussions of metaethics. Societies of rational agents need ethics to regulate their conduct. Out AIs aren't sophisticated enough to live in their own socieities. A wireheading AI isn't even going to be able to survive "in the wild". If you could build an artificial society of AI, then the questions of whether they spontaneously evolved ethics would be a very interesting and relevant datum. But AIs as we know them aren't good models for the kinds of entities to which morality is relevant. And Clippy is particularly exceptional example of an AI. So why do people keep saying "Ah, but Clippy..."...?

3Nornagest
Well, in this case it's because the post I was responding to mentioned Clippy a couple of times, so I thought it'd be worthwhile to mention how the little bugger fits into the overall picture of value stability. It's indeed somewhat tangential to the main point I was trying to make; paperclippers don't have anything to do with value drift (they're an example of a different failure mode in artificial ethics) and they're unlikely to evolve from a changing value system.
-4MugaSofer
Key word here being "societies". That is, not singletons. A lot of the discussion on metaethics here is implicitly aimed at FAI.

Isn't the idea of moral progress based on one reference frame being better than another?

0TheOtherDave
Yes, as typically understood the idea of moral progress is based on treating some reference frames as better than others.
0MugaSofer
No, because "better" is defined within a reference frame.

At last, an interesting reply!

[This comment is no longer endorsed by its author]Reply
6Nornagest
...with extremely low probability. It's far more likely that the Life field will stabilize around some relatively boring state, empty or with a few simple stable patterns. Similarly, a system subject to value drift seems likely to converge on boring attractors in value space (like wireheading, which indeed has turned out to be a problem with even weak self-modifying AI) rather than stable complex value systems. Paperclippism is not a boring attractor in this context, and a working fully reflective Clippy would need a solution to value drift, but humanlike values are not obviously so, either.
5TheOtherDave
Yes, value drift is the typical state for minds in our experience. Building a committed Clipper that cannot accidentally update its values when trying to do something else is only possible after the problem of value drift has been solved. A system that experiences value drift isn't a reliable Clipper, isn't a reliable good-thing-doer, isn't reliable at all. Next.
3DaFranker
I never claimed that it was controversial, nor that AIs didn't need to self-modify, nor that values are exempt. I'm claiming that updates and self modification do not imply a change of behavior towards behavior desired by humans. I can build a small toy program to illustrate, if that would help.
4nshepperd
The Futility of Emergence A paperclipper no more has a wall stopping it from updating into morality than my laptop has a wall stopping it from talking to me. My laptop doesn't talk to me because I didn't program it to. You do not update into pushing pebbles into prime-numbered heaps because you're not programmed to do so. Does a stone roll uphill on a whim? Perhaps you should study Reductionism first.
2DaFranker
Other key problem: Please unpack this and describe precisely, in algorithmic terms that I could read and write as a computer program given unlimited time and effort, this "ability to update" which you are referring to. I suspect that you are attributing Magical Powers From The Beyond to the word "update", and forgetting to consider that the ability to self-modify does not imply active actions to self-modify in any one particular way that unrelated data bits say would be "better", unless the action code explicitly looks for said data bits.

Arbitrary and biased are value judgments.

And they'ree built into rationality.

Whether more than one non-contradictory value system exists is the topic of the conversation, isn't it?

Non contradictoriness probably isn't a sufficient condition for truth.

-4TimS
Arbitrary and Bias are not defined properties in formal logic. The bare assertion that they are properties of rationality assumes the conclusion. Keep in mind that "rationality" has a multitude of meanings, and this community's usage of rationality is idiosyncratic. Sure, but the discussion is partially a search for other criteria to evaluate of the truth of moral propositions. Arbitrary is not such a criteria. If you were to taboo arbitrary, I strongly suspect you'd find moral propositions that are inconsistent with being values-neutral.
-2TimS
Arbitrary and biased are value judgments. If we decline to make any value judgments, I don't see any way to make those sorts of claims. Whether more than one non-contradictory value system exists is the topic of the conversation, isn't it?
4CCC
...I am surprised. I still can't imagine it myself, but I guess that means that someone can.

You are the monarch in that society, you do not need to guess which role you're being born into, you have that information. You don't need to make all the slaves happy to help your goals, you can just maximize your goals directly. You may choose any moral principle you want to govern your actions. The Categorical Imperative would not give you the best result.

For what value of "best"? If the CI is the correct theory of morality, it will necessarily give your the morally best result. Maybe your complaint is that it wouldn't maximise your persona... (read more)

-2MugaSofer
Not Beady, Anti-Clippy: an agent that is the precise opposite of Clippy. It wants to minimize the number of paperclips.
0Kawoomba
Not very rational for those to adopt a losing strategy (from their point of view), is it? Especially since they shouldn't reason from a point of "I could be the king". They aren't, and they know that. No reason to ignore that information, unless they believe in some universal reincarnation or somesuch. Yes. Which is why rational agents wouldn't just go and change/compromise their terminal values, or their ethical judgements (=no convergence). Starting out with different interests. A strong clippy accommodating a weak beady wouldn't be in its best self-interest. It could just employ a version of morality which is based on some tweaked axioms, yielding different results. There are possibly good reasons for us as a race to aspire to working together. There are none for a domineering Clippy to take our interests into account, yielding to any supposedly "correct" morality would strictly damage its own interests.

How many 5 year olds have the goal of Sitting Down WIth a Nice Cup of Tea?

2DaFranker
One less now that I'm not 5 years old anymore. Could you please make a real argument? You're almost being logically rude.
0MugaSofer
Why do you think adults sit down with a nice cup of tea? What purpose does it serve?
4MugaSofer
Biases are not determined by vote.

However, both the pebble-sorters and myself share one key weakness: we cannot examine ourselves from the outside; we can't see our own source code.

Being able to read all you source code could be ultimate in self-reflection (absent Loeb's theorem), but it doens't follow that those who can't read their source-code can;t self reflect at all. It's just imperfect, like everything else.

nshepperd110

This is where you are confused. Almost certainly it is not the only confusion. But here is one:

Values are not claims. Goals are not propositions. Dynamics are not beliefs.

A machine that maximises paperclips can believe all true propositions in the world, and go on maximising paperclips. Nothing compels it to act any differently. You expect that rational agents will eventually derive the true theorems of morality. Yes, they will. Along with the true theorems of everything else. It won't change their behaviour, unless they are built so as to send those actio... (read more)

-2MugaSofer
... why do you think people say "I don't like X, but I respect your right to do it"?

First of all, thanks for the comment. You have really motivated me to read and think about this more

That's what I like to hear!

If there are no agents to value something, intrinsically or extrinsically, then there is also nothing to act on those values. In the absence of agents to act, values are effectively meaningless. Therefore, I'm not convinced that there is objective truth in intrinsic or moral values.

But there is no need for morality in the absence of agents. When agents are there, values will be there, when agents are not there, the absence ... (read more)

But is the justification for its global applicability that "if everyone lived by that rule, average happiness would be maximized"?

Well, not, that's not Kant's justification!

That (or any other such consideration) itself is not a mandatory goal, but a chosen one.

Why would a rational agent choose unhappiness?

If you find yourself to be the worshipped god-king in some ancient Mesopotanian culture, there may be many more effective ways to make yourself happy, other than the Categorical Imperative.

Yes, but that wouldn't count as ethics. You ... (read more)

-2MugaSofer
If there are a lot of similar agents in similar positions, Kantian ethics works, no matter what their goals. For example, theft may appear to have positive expected value - assuming you're selfish - but it has positive expected value for lots of people, and if they all stole the economy would collapse. OTOH, if you are in an unusual position, the Categorical Imperative only has force if you take it as axiomatic. That's not a version of Kantian ethics, it's a hack for designing a society without privileging yourself. If you're selfish, it's a bad idea.
2Kawoomba
Scenario: 1) You wake up in a bright box of light, no memories. You are told you'll presently be born into an Absolute monarchy, your role randomly chosen. You may choose any moral principles that should govern that society. The Categorical Imperative would on average give you the best result. 2) You are the monarch in that society, you do not need to guess which role you're being born into, you have that information. You don't need to make all the slaves happy to help your goals, you can just maximize your goals directly. You may choose any moral principle you want to govern your actions. The Categorical Imperative would not give you the best result. A different scenario: Clippy and Anti-Clippy sit in a room. Why can they not agree on epistemic facts about the most accurate laws of physics and other Aumann-mandated agreements, yet then go out and each optimize/reshape the world according to their own goals? Why would that make them not rational? Lastly, whatever Kant's justification, why can you not optimize for a different principle - peak happiness versus average happiness, what makes any particular justifying principle correct across all - rational - agents. Here come my algae!

Yea, honestly I've never seen the exact distinction between goals which have an ethics-rating, and goals which do not

A number of criteria have been put forward. For instance, do as you would be done by. If you don't want to be murdered, murder is not an ethical goal.

My problem with "objectively correct ethics for all rational agents" is, you could say, where the compellingness of any particular system comes in. There is reason to believe an agent such as Clippy could exist, and its very existence would contradict some "'rational' corre

... (read more)
0Kawoomba
I'm not disputing that there are goals/ethics which may be best suited to take humanity along a certain trajectory, towards a previously defined goal (space exploration!). Given a different predefined goal, the optimal path there would often be different. Say, ruthless exploitation may have certain advantages in empire building, under certain circumstances. The Categorical Imperative in all its variants may be a decent system for humans (not that anyone really uses it). But is the justification for its global applicability that "if everyone lived by that rule, average happiness would be maximized"? That (or any other such consideration) itself is not a mandatory goal, but a chosen one. Choosing different criteria to maximize (e.g. noone less happy than x) would yield different rules, e.g. different from the Categorical Imperative. If you find yourself to be the worshipped god-king in some ancient Mesopotanian culture, there may be many more effective ways to make yourself happy, other than the Categorical Imperative. How can it still be said to be "correct"/optimal for the king, then? So I'm not saying there aren't useful ethical system (as judged in relation to some predefined course), but that because those various ultimate goals of various rational agents (happiness, paperclips, replicating yourself all over the universe) and associated optimal ethics vary, there cannot be one system that optimizes for all conceivable goals. My argument against moral realism and assorted is that if you had an axiomatic system from which it followed that strawberry is the best flavor of ice cream, but other agents which are just as intelligent with just as much optimizing power could use different axiomatic systems leading to different conclusions, how could one such system possibly be taken to be globally correct and compelling-to-adopt across agents with different goals? Gandhi wouldn't take a pill which may transform him into a murderer. Clippy would not willingly modify i

Spun off from what, and how?

I am not sure I can expalin that succintly at the moment. It is also hard to summarise how you get from counting apples to transfinite numbers.

Why does their being rational demand that they have values in common? Being rational means that they necessarily share a common process, namely rationality, but that process can be used to optimize many different, mutually contradictory things. Why should their values converge?

Rationality is not an automatic process, it is skill that has to be learnt and consciously applied. Indivi... (read more)

Yes, but the fact that the universe itself seems to adhere to the logical systems by which we construct mathematics gives credence to the idea that the logical systems are fundamental, something we've discovered rather than producing. We judge claims about nonobserved mathematical constructs like transfinites according to those systems,

But claims about transfinities don't correspond directly to any object. Maths is "spun off" from other facts, on your view. So, by analogy, moral realism could be "spun off" without needing any Form of... (read more)

0Desrtopa
Spun off from what, and how? Speaking as a utilitarian, yes, utilitarianism does care about what values are. If I value paperclips, I assign utility to paperclips, if I don't, I don't. Why does their being rational demand that they have values in common? Being rational means that they necessarily share a common process, namely rationality, but that process can be used to optimize many different, mutually contradictory things. Why should their values converge? So what if a paperclipper arrives at "maximize group utility," and the only relevant member of the group which shares its conception of utility is itself, and its only basis for measuring utility is paperclips? The fact that it shares the principle of maximizing utility doesn't demand any overlap of end-goal with other utility maximizers. But, as I've pointed out previously, intuitions are often unhelpful, or even actively misleading, with respect to locating the truth. If our axioms are grounded in our intuitions, then entities which don't share our intuitions will not share our axioms. No, but neither do I, so I don't see why that's relevant.

Since no claim has a probability of 1.0, I only need to argue that a clear majority of rational minds converge.

3Desrtopa
Yes, but the fact that the universe itself seems to adhere to the logical systems by which we construct mathematics gives credence to the idea that the logical systems are fundamental, something we've discovered rather than producing. We judge claims about nonobserved mathematical constructs like transfinites according to those systems, But utility is a function of values. A paperclipper will produce utility according to different values than a human. Why would most rational minds converge on values? Most human minds converge on some values, but we share almost all our evolutionary history and brain structure. The fact that most humans converge on certain values is no more indicative of rational minds in general doing so than the fact that most humans have two hands is indicative of most possible intelligent species converging on having two hands. It means we should be aware of what our intuitions are and what they've developed to be good for. Intuitions are evolved heuristics, not a priori truth generators. It seems like you're equating intuitions with axioms here. We can (and should) recognize that our intuitions are frequently unhelpful at guiding us to he truth, without throwing out all axioms. If I did, I don't remember them. I may have, I may have felt someone else adequately addressed them, I may not have felt it was worth the bother. It seems to me that you're trying to foist onto me the effort of locating something which you were the one to testify was there in the first place. And philosophers frequently fall into the pattern of believing that other philosophers disagree with each other due to failure to understand the problems they're dealing with. In any case, I reject the notion that dismissing large contingents of philosophers as lacking in competence is a valuable piece of evidence with respect t crankishness, and if you want to convince me that I am taking a crankish attitude, you'll need to offer some other evidence.

I haven't seen anything to say that is for meta discussion, it mostly isn't de facto, and I haven't seen a "take it elsewhere" notice anywhere as an aternative to downvote and delete.

All that's needed to is reject the idea that there are some mysterious properties to sensation which somehow violate basic logic and the principles of information theory.

Blatant strawman.

Banning all meta discussion on LW of any kind seems like an increasingly good idea - in terms of it being healthy for the community, or rather, meta of any kind being unhealth

Have you considered having a separate "place" for it?

2[anonymous]
http://lesswrong.com/lw/gkv/official_lw_uncensored_thread_on_reddit/

But being able to handle criticism properly is a very important rational skill. Those who feel they cannot do it need to adjust their levels of self-advertisement as rationalists accordingly.

5Richard_Kennaway
You are absolutely right. Some parts of this very important rational skill are: properly discerning genuine criticism from trolling; properly discerning whether the person posting it is a useful or a harmful presence in the forum; properly deciding a useful course of action. I think that Eliezer has indeed demonstrated possession of this very important rational skill in his handling of V_V's criticism.

Maths isn't very relevant to Rand's philosophy. What's more relevant about her Aristoteleanism is her attitude to modern science; she was fairly ignorant. and fairly sceptical, of evolution, QM, and relativity.

Is unpleasantness the only criterion? Nobody much likes criticism, but it is hardly rational to disregard it becuase you don't like it.

I should've said, "updatable terminal goals".

You can make the evidence compatble with the theory of terminal values, but there is still no support for the theory of terminal values.

0Bugmaster
I personally don't know of any evidence in favor of terminal values, so I do agree with you there. Still, it makes a nice thought experiment: could we create an agent possessed of general intelligence and the ability to self-modify, and then hardcode it with terminal values ? My answer would be, "no", but I could be wrong. That said, I don't believe that there exists any kind of a universally applicable moral system, either.

a successful rationalist organisation should be right up at the zero end of the scale

Because everyone knows that reversed stupidity is the best form of rationality ever.

Here are some guidelines for the new ultra-rational community to follow:

  • Don't have any leadership. If someone tries to organize something, make sure you criticize them loudly and question their motives, until they crawl away crying.
  • Prevent new people from joining you.
  • Money making or any success in real life should be considered shameful.
  • Emphasise that there is no truth, no rea
... (read more)
7hairyfigment
If you tell me that all successful organization do X, and then advise me not to do X, I'll start to doubt that you have my best interests at heart. At least if I can think of several defunct clubs/ideas from my own experience that did not do X (which I think I can).
1Morendil
Insofar as many corporations would check more items from that list than I suspect the Boston LW group would, yes. Insofar as many of the items are vague enough to apply to any social group that elicits loyalty from its members, yes. One problem is relative terms like "excessive" in "excessively zealous, unquestioning commitment". What observations, precisely, count as indications of "excessive" behavior in this regard? Or "preoccupied with making money" - well who isn't? Again, what's a cult-indicative level of preoccupation? It's going to be hard to beat, e.g. the startup community in terms of being obsessed with money, so this indicator totally fails to discriminate cults in any useful manner. If you said "cults assert and enforce an exclusive and all-encompassing claim to members' or prospective members' income and wealth", that would be more diagnostic. (But then you couldn't arbitrarily designate any group you didn't like as being a cult. Oh well.)
9Vaniver
Really? I get much more of the "have a deliberately built, internally consistent, and concordant with reality worldview" vibe from EY and LW than I do from most of the new atheist movement. Have you read the Death Spirals and the Cult Attractor sequence?
-2MaoShan
No need, you're doing a fine job of that all by yourself.

" values are nothing to do with rationality"=the Orthogonality Thesis, so it's a step in the argument.

2Paul Crowley
It feels to me like the Orthogonality Thesis is a fairly precise statement, and moral anti-realism is a harder to make precise but at least well understood statement, and "values are nothing to do with rationality" is something rather vague that could mean either of those things or something else.

Too see why someone might think that, imagine the following scenario: You find scientific evidence for the fact that if one forces the minority of the best-looking young women of a society at gunpoint to be of sexual service to whomever wishes to be pleased (there will be a government office regulating this) increases the average happiness of the country.

If you disregard the happiness of the women, anyway

In other words, my argument questions that the happiness (needs/wishes/etc.) of a majority is at all relevant. This position is also known as individ

... (read more)
4jt4242
No, it suffices if less women's happiness sacrificed are needed than the amount of men whose happiness will be increased (assuming the "amount of happiness" - whatever that is to mean in the first place - is equal per individual). Then you can regard the happiness of women and still score a net increase in happiness. That's the whole point of the argument. I don't understand what you were saying in the second sentence.
Load More