* one who reliably wins
* one who can reliably be expected to speak truth
Yeah it must have been getting pretty bad if we got to the point that the word could mean two contradictory things. ;-)
A good reason to take this suggestion to heart: The terms "rationality" and "rational" have a strong positive value for most participants here—stronger, I think, than the value we attach to words like "truth-seeking" or "winning." This distorts discussion and argument; we push overhard to assert that things we like or advocate are "rational" in part because it feels good to associate our ideas with the pretty word.
If you particularize the conversation—i.e., you are likely to get more money by one-boxing on Newcomb's problem, or you are likely to hold more accurate beliefs if you update your probability estimates based solely on the disagreement of informed others—than it is less likely that you will grow overattached to particular procedures of analysis that you have previously given an attractive label.
Definitions of words usually only make much sense in some restricted part of reality, where certain things correlate with certain other things.
For "rationality", it is the case that a certain collection of human mental habits tend to correlate with success in a wide field of endeavors, and that the primary reason for this is that the human mind is an example of a somewhat broken truth seeker (i.e. a partially accurate map), and in many cases mending the map causes the human in question to better achieve his/her goals.
However, there are some contexts in which "mending" the map does not cause the human to better achieve his/her goals. These situations (such as the man who would do best to assign probability zero to his wife cheating on him, for example) reflect the fact that the human mind is not a perfect disembodied AGI. We leak our preferences out in our body language and mannerisms. We have preferences that supervene on our own mental states (making the map a self-referential one). We find it hard to motivate ourselves to act, and sometimes believing that things are better than they really are can motivate us better.
It is thus the case that rationality-as-winning can be very weird in general. There are possible worlds where dark-side epistemology is good for you. Imagine an impoverished medieval peasant being confronted with the truth about his existence - religion was probably the only thing that made life even remotely worth living. It is only the empirical content of this world that makes rationality-as-winning look a lot like rationality-as-accurate-map, and thus allows us to further our goals in a particularly general and simple way. Eliezer's observations about dark side epistemology and reality as an interconnected web, and about how if you abandon rationality-as-accurate-map because you think it will be good for you (i.e. you delude yourself), you can never be sure that there isn't something crucial you've missed - something like cryonics - that will change the entire calculation. This tells me that once the optimal policy on self-delusion looks (naively) like "have an almost accurate map", then the true optimal policy (taking black swans like cryonics into account) is to have a completely accurate map. Thus the shift from a world where map-accuracy and winningness are not the same thing to one where they're identical a posteriori is a nonlinear shift, and in my mind this legitimizes the use of the term "rationality".
It is my opinion that the balance towards rationality-as-winning being rationality-as-accurate-map has only just tipped. The possibility of a technological singularity, of technological immortality, of technological utopia and the insurance that cryonics offers really do tip the balance I think. Without these, even in a comfortable world like the world of the US/EU today, most people's rationality-as-winning probably involves a lot of irrationality-as-inaccurate-map and dark side epistemology to perform the cover up.
What a ... convenient coincidence ... that we live just past the transition where perfect rationality becomes the optimal strategy. Doesn't it seem a little too convenient?
Yes, actually, I had the same thought myself last night and was about to edit.
Now, as a matter of fact it does seem likely to me that the turnaround was between 1950 and today. This does seem like a "too good to be true" coincidence... suggesting the alternative hypothesis that I am busy rationalizing.
But I've reviewed the object-level arguments, and they seem solid.
I think that the explanation for half of the co-incidence is that science is a common cause of rationality arts and the improved standard of living that makes them worth applying. JulianMorrison simultaneously made the same suggestion.
The other half of the co-incidence (why aren't we alive waaaaay after it becomes rational to be rational) is actually not such a co-incidence... there is nothing interesting to be learned by observing that we're alive just after technology becomes quite an important factor. (other than the usual doomsday argument/negative singularity stuff which we won't go into)
EDIT: the possibility that we are rationalizing should cause us to decrease our confidence somewhat in the position stated above. Though, the fact that JulianMorrison thought of the same explanation independently of me should be noted. We should now consider more seriously the alternative hypothesis that "perfect rationality" is still not optimal, and that deliberately deluding ourselves in some significant part of our lives is a good idea. [I stand by my position that there were definitely some periods where rationality was a real downer]. Thanks to gjm and johnicholas for pointing this out.
I voted it down, and this is my reasoning:
This point does not seem to degrade the comment.
Rationalization is the standard method of generating explanations - first, one acts, then one comes up with a set of reasons why that happened, which are not necessarily causally related to the action. Wish I had a link to the relevant studies handy.
it's ridiculous to expect a comment to even mention the controversy around the Singularity and the 'anthropic argument'. If you don't know what they are, then you can look them up in all their controversial glory. If you do, then you already know they're controversial. And if something is 'nearly off-topic' then by the definition of 'nearly' it's not off-topic, so I'm not sure what your point is there. And to find a singularity-fan in this crowd should not be a surprise to you; it's another perspective from which to approach the question, and it's clear that this is the perspective of the poster.
From About Less Wrong:
To prevent topic drift while this community blog is being established, please avoid mention of the following topics on Less Wrong until the end of April 2009:
Nick Bostrom's introduction to the Doomsday Argument is an example of smart, cautious discussion of anthropic reasoning.
You should take the fact that the best argument that you can find for the proposition: "Rationality is optimal now, but it wasn't in 1950." is appealing to the Doomsday Argument, as evidence that your brain is in rationalization mode.
But ... (and now I'm genuinely curious) why aren't we living in a period way after rationality became the optimal choice? JulianMorrison and my suggestion provides the lower bound, but what is the upper bound?
To falsify the conjunction "Rationality is optimal now" and "Rationality was not optimal previously", you only need to falsify one of the conjuncts. For example, "Rationality is not optimal now" or "Rationality was optimal previously".
EDIT: I said that awkwardly. To change your mind regarding "Rationality is optimal now and rationality was not optimal previously", you would have to change your mind regarding one of the conjuncts. For example, you could accept the statement "Rationality is not optimal now."
Robin Hanson has posted on the costs of rationality.
So, anthropic reasoning involves using facts about how the observer came into being to “explain” certain supposed coincidences and thereby not give so much weight to alternative hypotheses which might need to be invoked to explain the coincidence.
In this case, there is a coincidence between us asserting that rationality is good for us, and us being the first generation out of a long line of humans for whom this is the case. (and, indeed, the same argument applies spatially as temporally; rationality is probably a bad move for many very disadvantaged people in the world today).
The alternative hypothesis under consideration is “rationality is not good for you, you are just rationalizing”.
So, I assume that I am sampled from the set of people who ask the question "is it optimal to be rational, or to delude myself?". What is the probability of me answering "yes"? Well, JulianMorrison argues (correctly, IMO) that there is a systematic correlation between being able to ask the question and answering "yes", so the probability is not worryingly small. Nothing unusual has happened here.
So we should not be suspicious that we are rationalizing just because we answered "yes".
Secondly, what is the probability of me finding myself to be the first (or second) generation of humans for which the answer to this question is "yes"? In the case where there are zillions of similar humans in the future, this probability could be very small. But... there's no interesting alternative hypothesis to explain this coincidence, so we can't conclude anything particularly interesting.
I'd love to know why someone downvoted this...
(this comment also downvoted to 0)
ROFL someone is out to get me, I can see ;-0
Aaah! negative karma! Everybody hates me! I'm considering killing myself if this comment gets downvoted any more...
Thus the shift from a world where map-accuracy and winningness are not the same thing to one where they're identical a posteriori is a nonlinear shift, and in my mind this legitimizes the use of the term "rationality".
Where's the nonlinearity, and why does nonlinearity legitimize the term "rationality"?
The nonlinearity is the following: once you realize the dark side epistemology/interconnected web of reality material, and you already think that the most "winning" course of action is to have an almost accurate map, you should decide that the most "winning" way is to have a fully accurate map. The intuition is that it is not winningness-promoting to deliberately make a small subset of your beliefs inaccurate.
This legitimizes the term because it is then empriically the case that winningness and map accuracy coincide exactly, so we can afford to use them interchangeably.
EDIT:
I think that it is fair to use the term rationality most of the time. In certain discussions, such as this one, it may be neccessary to resort to "this behavior will track truth better" and "this behavior will make you win more"
I voted this up, just for the line "It is my opinion that the balance towards rationality-as-winning being rationality-as-accurate-map has only just tipped." This is plausible and interesting.
Human psychology is so weird that having correct beliefs actually works against winning reliably. For example, having correct beliefs requires you to make a special effort to seek evidence that your beliefs are wrong, making you less certain than others and therefore less determined.
Re: Are there any others?
Well, yes. Rationality, the way I think of it, isn't about "winning" or "truth-seeking".
If you think of a cybernetic diagram of an organism - like this:
Sensory input -> Computations -> Motor output
...then I think "rationality" needs to be confined to the middle unit. It is a computational process. You might need some motor output in order to be able to detect it - but that isn't part of rationality itself.
Truth-seeking is a goal. It is one goal among the many possible that it is possible to rationally pursue. Rational agents often adopt truth-seeking as a proximate goal - whatever they want to do - but embracing false beliefs is sometimes rational too - it depends on what your goal is.
For me, rationality has a lot to do with the valid use of inductive and deductive reasoning in pursuit of a goal.
I think "probability updates under Bayes' rule" is very clever and highly accurate, and it gets to just what you're talking about. Also, since this thread is trending towards everyone defining (or at least characterizing) rationality for themselves, here goes: rationality is what happens when evidence is recognized by a consciousness, subjected to ordered thought, and used to form or modify beliefs.
That's as close as I can get to "correct" for myself with a few minutes of thought and natural language. It seems to fit with the notion of rationality as a computational process.
I think it's fine to use the terms if you say what you mean by them, and not especially bad to use them even if you don't. We can't define all our terms. Why single those ones out as problematic? More potential for confusion lurks in the words we're not watching as closely.
I will not cease using perfectly good words.
I will, however, ask that people be prepared to define and explain the words when they use them. Words are problematic only when they become empty signifiers, labels attached to nothing.
Words are problematic only when they become empty signifiers, labels attached to nothing.
this is precisely what I'm worried about -- that eventually we'll be using "rationality" to mean "things that LW readers like"
Y'know how we get around that? Insist on definitions. They're still pretty sparse on the ground, here. And the one that's had the most publicity is a very poor match for the generally-accepted meaning of the term.
Normally, people manage to communicate using our informal, muddly, complicated, natural language abilities. Sometimes this breaks down when we're discussing value-laden or highly abstract concepts.
Breaking words down into definitions doesn't solve the problem - the components that you define with need to be communicated, too. This lowest-level communication needs to be informal, non-defined primitives.
Tabooing words reboots the informal process of achieving communication, without the fuss of arguing about whether a definition is correct, or queries about which definition you are using.
"Normally, people manage to communicate using our informal, muddly, complicated, natural language abilities."
I think that, in actuality, they don't. Or rather, they communicate very little: mostly by indicating positions that the listener is already familiar with.
Ever try explaining a truly new idea to someone? With most people, I find that if they don't already have a referent, they simply can't understand, because they're not used to extracting complex information from natural language.
We're in agreement. The position that I was arguing against is something like: "People can't communicate unless they first define their terms." That would be an infinite regress; the only possibility would be that people never manage to communicate.
Okay, I'll accept that.
I offer a restatement: people can't communicate at a complex and abstract level unless their words are first defined in terms of words with already-accepted and -understood meanings.
If I begin to talk about gilxorfibbin without explaining what that is, it's unlikely the context will make it possible for you to know what I'm discussing.
The problem is that definitions are not hierarchical, you never get to the lowest level, because there isn't one. You need to choose a way to the target concept that communicates it as unambiguously as possible. The words spoken by one person guide another on his own map, pointing to the deeper and deeper concepts that require nontrivial arrangements from the words to single out, or even build anew.
Some words are broken, and lead the listener in the swamps. We should avoid these words, and use other healthier landmarks instead. Sometimes it requires a lengthy detour to get around the swamps, but the road is not necessarily any bumpier or conversely more streamlined than what would be expected of the original one.
Words can become less useful when they attach to too much as well as too little. A perfectly drawn map that indicates only the position and exact shape of North America will often be less useful than a less-accurate map that gives the approximate location of its major roads and cities. Similarly, a very clearly drawn map that does not correspond to the territory it describes is useless. So defining terms clearly is only one part of the battle in crafting good arguments; you also need terms that map well onto the actual territory and that do so at a useful level of generality.
The problem with the term "rationality" isn't that no one knows what it means; there seems to be wide agreement on a number of tokens of rational behavior and a number of tokens if irrational behavior. Rather, the problem is that the term is so unspecific and so emotionally loaded that it obstructs rather than furthers discussion.
I think we should be careful of elegant variation, which can be awkward and introduce ambiguity. Rather than simply using sobriquets or synonyms like "truth-seeking", "lucidity", or "ratiocination", we might do better to interrogate each other on what rationality is, and make frequent, almost-repetitive reference to the essentials of rationality. This would work especially well for people with idiosyncratic definitions.
There was a (tiny) movement about 20 years ago to get people to stop using the word "is". Usually, an "is" renders a judgement while concealing the reasons: "Cindy is sweet. The GPL is stupid and destructive."
I think I can make as good a case for banning "is" as for banning "rationality". And if we should ban "is", what shouldn't we ban? Can you name any words that shouldn't be banned?
Maybe we should just point.
There was a (tiny) movement about 20 years ago to get people to stop using the word "is". Usually, an "is" renders a judgement while concealing the reasons: "Cindy is sweet. The GPL is stupid and destructive."
I think I can make as good a case for banning "is" as for banning "rationality". And if we should ban "is", what shouldn't we ban? Can you name any words that shouldn't be banned?
I find your arguments a bit muddled and confusing: I can make as good a case for banning genocide as I can for banning pleasure (e.g. by making a very poor case in both cases). That doesn't mean I've established that either one should be banned; nor does it mean that I've established that they are equally "ban-worthy"; nor have I established that the reasons for banning one are in any way related with the reasons for banning the other.
It seems like your arguments for banning "is" is that it could be used to "renders a judgement while concealing the reasons". But if people think it's appropriate to render judgment without concealing reasons, then there's no reason to ban "is", correct?
Contrast this with the argument for banning "rational" in that people here are using it to mean different thing, and we're having a lot of confusion due to not knowing which meaning is intended.
Even if we accept that both arguments are equally logically sound, we might choose to ban one without banning the other based on our values (e.g. if we very highly value the non-concealing of reasons, but don't value lack of confusion, we may choose to ban "is" without banning "rational").
Are there any others?
Eliezer's not saying the obvious so I will...
Hitting small targets in large search spaces to produce coherent real-world effects that steer reality into regions that are higher in your preference ordering.
That's exactly what tabooing "rationality" does--with the added benefit of bringing all definitions out into the open. The conventional definition of rationality should be as explicit as idiosyncratic definitions. Furthermore, people in general are bad at noticing when they have idiosyncratic ideas and tend to assume everyone uses words in the same way they do.
I agree with this point generally but it is difficult to find specific examples because they will be heavily context-dependent. Ratiocinative is one probably underused word, as is lucid.
Rationalist: one who seeks to only have contagious beliefs, that is, beliefs which are tightly correlated/entangled with reality.
Rationalist: one who seeks to only have contagious beliefs, that is, beliefs which are tightly correlated/entangled with reality.
Note that while I agree with Eliezer that "rational beliefs are contagious", I disagree with the claim that "contagious beliefs are rational". See, e.g., religion.
Error reduction?
Probability or possibility pruning?
Parachute venting? (The mechanism for letting hot air out of an air balloon to bring it back to the ground).
Related on OB: Taboo Your Words
I realize this seems odd on a blog about rationality, but I'd like to strongly suggest that commenters make an effort to avoid using the words "rational," "rationality," or "rationalist" when other phrases will do. I think we've been stretching the words to cover too much meaning, and it's starting to show.
Here are some suggested substitutions to start you off.
Rationality:
Rationalist:
Are there any others?