You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Is there a rationalist skill tree yet?

15 Post author: fowlertm 30 January 2015 04:02PM

A while back I came across a delightful web developer skill tree, and I was wondering if technical rationality has gotten to the point where someone could make one of these for an aspiring rationalist.

I think seeing a clear progression from beginning skills to advanced ones laid out graphically helps those starting on the path conceptualize the process. 

Comments (14)

Comment author: [deleted] 30 January 2015 09:38:48PM 13 points [-]

Here's a decent one that's two levels deep: http://lesswrong.com/lw/fc3/checklist_of_rationality_habits/

Comment author: John_Maxwell_IV 31 January 2015 07:07:25AM 7 points [-]

Previously on LW: 1, 2, 3.

Comment author: redding 31 January 2015 12:38:21AM 3 points [-]

I think such a tree would depend in large part on what approach one wants to take. Do you want to learn probability to get a formal foundation of probabilistic reasoning? As far as I know, no other rationality skill is required to do this, but a good grasp of mathematics is. On the other hand, very few of the posts in the main sequences (http://wiki.lesswrong.com/wiki/Sequences#Major_Sequences) require probability theory to understand. So, in a sense, there is very little cross-dependency between mathematical understanding of probability and the rationality taught here. On the other hand, so many of the ideas are founded on probability theory, that it seems odd that they wouldn't be required. Thoughts?

Comment author: MathiasZaman 30 January 2015 06:22:13PM 3 points [-]

I've never seen one, but it'd be a great resource to have especially if it links to other resources that teach certain skills.

I do, however, think that the field of rationality isn't codified enough to make a meaningful skill-tree. You can probably make a simple or bare-bones version and I think having that would be better than having nothing.

What would be on it? Skepticism, akrasia-fighting, research, social skills...

Comment author: michael_b 31 January 2015 10:29:54AM *  2 points [-]

Great request. Along a similar line, what about a decision tree for evaluating claims?

  1. Are you being asked to think about past decisions, prior beliefs, or intent? Take steps to rule out hindsight bias.

  2. Does the claim challenge prevailing beliefs, possibly alleging conspiracy? Consider confirmation bias.

...

(sorry if this is a FAQ)

Comment author: spriteless 30 January 2015 08:38:09PM 2 points [-]

Ancedotal Evidence suggests that the first, most important skill, is being able to admit you are wrong. Taken to far though, and it results in an useless humble platitudes. Paired with being able to look at the universe around you to find what is right, I think it is enough enough to recreate everything. I would go so far as to say that Bayes' Theorem is just a mathematical formalization of those two ideas.

Comment author: imuli 30 January 2015 06:38:24PM 2 points [-]

There is a dependency tree for Eliezer Yudkowsky's early posts. It's not terribly pretty, but with a couple hours and a decent data presentation toolkit someone could probably make a pretty graphical version. It doesn't include a lot of later contributions by other people, but it'd be a start.

Comment author: [deleted] 30 January 2015 07:45:44PM 2 points [-]

I'm not sure that's the same as a skill tree.

Comment author: fowlertm 31 January 2015 02:27:21PM 0 points [-]

I thought of that as well, it does need some work done in terms of presentation. It'd be a good place to start, yes.

Comment author: Ishaan 31 January 2015 03:14:13AM *  1 point [-]

A good starting point would be to have people brainstorm examples of things that they had been exposed to but did not and could not understand at first, but which clicked together after they saw something else.

When I mentally map useful concepts talked about round these parts, they fall naturally into philosophy, metacognition, and a "practical" category composed of habits, skills, and knowledge. However, to the extent that concepts benefit from being known in order, they don't usually have more than 1 or 2 levels. I guess philosophy comes closest to having a "tree" of sorts.

For example, look at this. It is very appealing to the intuition. It sets up a progression of insight which is sort of valid, maybe...but...does the 1-2 emotional transition really preclude realizations 3 and 4? Does Realization 3 actually help with Step 2 behavior as claimed? Are there not people who convincingly proclaim step 4 while not really getting step 3? Maybe these are actually 3 entirely separate things?

It's not trivial to know how things are related.

Comment author: LawrenceC 01 February 2015 05:45:36PM *  1 point [-]

I tried making one just for the math behind rationality/decision theory back in October, but I never got around to finishing it. The main problems I ran into were:

  • Where should the skill tree start? I'm sure that basic math like algebra, geometry, trig, etc are all really useful, but I'm not sure about the dependencies between them. I ended up lumping them all into "basic mathematics".

  • How should the skill tree split subjects? Many subjects are best learned iteratively - for example, it's probably best to get a rudimentary understanding of probability theory, then learn more probability theory later on once you've picked up other related subjects (Linear Algebra, Multivariate Calculus, etc) and then again after more subjects (Measure theory). The complication is that these other subjects are often split into different "levels". I found that I didn't have enough familiarity with math to split subjects naturally.

One method that seems promising is taking a bunch of textbooks/courses, and trying to figure out the dependencies between them.

Comment author: fowlertm 03 February 2015 05:28:36PM 1 point [-]

Agreed. I think in light of the fact that a lot of this stuff is learned iteratively you'd want to unpack 'basic mathematics'. I'm not sure of the best way to graphically represent iterative learning, but maybe you could have arrows going back to certain subjects, or you could have 'statistics round II' as one of nodes in the network.

It seems like insights are what you're really aiming at, so maybe instead of 'probability theory' you have a node for 'distributions' and 'variance' at some early point in the tree then later you have 'Bayesian v. Frequentist reasoning'.

This would help also help you unpack basic mathematics, though I don't know much about the dependencies either. I hope too, soon :)

Comment author: Coscott 31 January 2015 05:36:02PM -2 points [-]

Rationality skills are not something you can complete and move on to the next level. If rationality moves into your system 1, then you are doing it wrong (or maybe doing it REALLY REALLY well).

Comment author: imuli 31 January 2015 07:31:24PM 3 points [-]

Noticing when you're confused and confidence calibration are two rationality skills that are necessary to have in your system 1 in order to progress as a rationalist… and much of instrumental rationality can be construed as retraining system 1.