I don't think there's a single defining point of difference, but I tend to think of it as the difference between the traditional social standard of having beliefs you can defend and the stricter individual standard of trying to believe as accurately as possible.
The How to Have a Rational Discussion flowchart is a great example of the former: the question addressed there is whether you are playing by the rules of the game. If you are playing by the rules and can defend your beliefs, great, you're OK! This is how we are built to reason.
X-rationality emphasizes having accurate beliefs over having defensible beliefs. If you fail to achieve a correct answer, it is futile to protest that you acted with propriety. Instead of asking "does this evidence allow me to keep my belief or oblige me to give it up?", it asks "what is the correct level of confidence for me to have in this idea given this new evidence?"
Eliezer uses "Traditional Rationality" to mean something like "Rationality, as practised by scientists everywhere, especially the ones who read Feynman and Popper". It refers to the rules that scientists follow.
A surely incomplete list of deficiencies:
In some ways, Eliezer is too hard on Traditional Rationalists (TRists). In the "wild and reckless youth" essay, which you cite, he focuses on how TR didn't keep him from privileging a hypothesis and wasting years of his life on it.
But TR, as represented by people like Sagan and Feynman, does enjoin you to believe things only on the basis of good evidence. Eliezer makes it sound like you can believe whatever crazy hypothesis you want, as long as it's naturalistic and in-principle-falsifiable, and as long as you don't expect others to be convinced until you deliver good evidence. But there are plenty of TRists who would say that you ought not to be convinced yourself until your evidence is strong.
However, Eliezer still makes a very good point. This injunction doesn't get you very far if you don't know the right way to evaluate evidence as "strong", or if you don't have a systematic method for synthesizing all the different evidences to arrive at your conclusion. This is where TR falls down. It gives you an injunction, but it leaves too much of the details of how to fulfill the injunction up to gut instinct. So, Eliezer will be contributing something very va...
I just started listening to THIS (perhaps 15min of it on my drive to work this morning), and EY has already mentioned a little about traditional rationality vs. where he is now with respect to reading Feynman. I'm not sure if he'll talk more about this, but Luke's page does have as a bullet point of the things covered:
Eliezer’s journey from ‘traditional rationality’ to ‘technical rationality’
so perhaps he'll continue in detail about this. Off hand, all I can specifically remember is that at one point he encountered some who thought that multiple routes...
One relevant attempt at a definition:
I will be using "extreme rationality" or "x-rationality" in the sense of "techniques and theories from Overcoming Bias, Less Wrong, or similar deliberate formal rationality study programs, above and beyond the standard level of rationality possessed by an intelligent science-literate person without formal rationalist training."
In one essay, Eliezer seems to be saying that Traditional Rationality was too concerned with process, whereas it should have been concerned with winning. In other passages, it seems that the missing ingredient in the traditional version was Bayesianism (a la Jaynes). Or sometimes, the missing ingredient seems to be an understanding of biases (a la Kahneman and Tversky).
All of those are problems with traditional rationality, and Elizeer has critiques traditional rationality for all of them. Traditional rationality should have helped Elizeer more than i...
I actually don't care about kharma - I'm not posting to get good kharma. Neither is curi. Disagreements should be resolved by discussion and by criticism, not by voting. I was just wondering how many people who disagree with Bayesianism end up with 0 kharma on LW and whether that isn't a bias? BTW, how do you know the reason something got downvoted?
With regard to your comments:
I have not found something on LW arguing that induction is impossible, the Popperian position. I have read a bunch of stuff here (done some homework) and it seems to me to be in the inductivist tradition of Aristotelian philosophy. I know other people who say the same thing and LW'ers that I have talked to seem incredulous that induction is impossible. So if you claim not to be in this mainstream tradition, I don't see how that can be and asking for material I cannot find is reasonable.
That wasn't an attempt to get upvotes. It was a comment to curi, who I know.
If I just commented on the first quote, people would have accused me of disputing the definition (which they did anyway - oh well). The "rules followed by scientists" refers to "traditional philosophy", by which Eliezer/Oscar mean Popper. Some commenters think Eliezer is only criticizing pop-culture. That is not so: he is criticizing Popper, and there are other posts where he makes this explicit. So Popper has everything to do with this.
You said not to start any replies with "lol". Popperians will try doing different things in conversation to see how the other person reacts. Are they concerned with style over substance? Do they place too much emphasis on emotional reactions? Are they conformists? I wasn't doing that in this instance, but by enforcing rigid standards of communication you lose knowledge. curi talks more about this in his threads.
I actually don't care about kharma - I'm not posting to get good kharma. Neither is curi. Disagreements should be resolved by discussion and by criticism, not by voting. I was just wondering how many people who disagree with Bayesianism end up with 0 kharma on LW and whether that isn't a bias? BTW, how do you know the reason something got downvoted?
This reads to me as "I don't care about karma, just about knowledge that can be derived from karma." These two positions seem to be, for all practical purposes, indistinguishable.
Also, for #1, AFAIK bayesians do not seek knowledge in the platonic sense.
In several places in the sequences, Eliezer writes condescendingly about "Traditional Rationality". The impression given is that Traditional Rationality was OK in its day, but that today we have better varieties of rationality available.
That is fine, except that it is unclear to me just what the traditional kind of rationality included, and it is also unclear just what it failed to include. In one essay, Eliezer seems to be saying that Traditional Rationality was too concerned with process, whereas it should have been concerned with winning. In other passages, it seems that the missing ingredient in the traditional version was Bayesianism (a la Jaynes). Or sometimes, the missing ingredient seems to be an understanding of biases (a la Kahneman and Tversky).
In this essay, Eliezer laments that being a traditional rationalist was not enough to keep him from devising a Mysterious Answer to a mysterious question. That puzzles me because I would have thought that traditional ideas from Peirce, Popper, and Korzybski would have been sufficient to avoid that error. So apparently I fail to understand either what a Mysterious Answer is or just how weak the traditional form of rationality actually is.
Can anyone help to clarify this? By "Traditional Rationality", does Eliezer mean to designate a particular collection of ideas, or does he use it more loosely to indicate any thinking that is not quite up to his level?