This is the post that finally prompted me to make an account on the site; I've been lurking and reading for quite some time, but while I haven't felt (most of) the topics to be over my head, I also did not think my (sometimes sparse) background in the subjects gave me much to add.
In the past I've been struck by how useful some of the articles here tend to be, and the simultaneous regret that they would not be likely to reach the people they would assist the most. I think that this post provides an excellent breakdown of ways to reach more people more effectively, and I hope that in the near future I'll be able to contribute to that goal. (And in the spirit of naming specifics, I can hopefully provide some biology- and evolution-centered tutorials, and assuming I can keep my other creative writing projects moving as well, I might very well try my hand at the story approach. I do ADORE HPatMoR.)
I look forward to what you put together!
My deconversion from Christianity had a large positive impact on my life. I suspect it had a small positive impact on the world, too. (For example, I no longer condemn gays or waste time and money on a relationship with an imaginary friend.) And my deconversion did not happen because I came to understand the Bayesian concept of evidence or Kolmogorov complexity or Solomonoff induction. I deconverted because I encountered some very basic arguments for non-belief, for example those in Dan Barker's Losing Faith in Faith.
Less Wrong has at least two goals. One goal is to raise the sanity waterline. If most people understood just the basics Occam's razor, what constitutes evidence and why, general trends of science, reductionism, and cognitive biases, the world would be greatly improved. Yudkowsky's upcoming books are aimed at this first goal of raising the sanity waterline. So are most of the sequences. So are learning-friendly posts like References & Resources for LessWrong.
A second goal is to attract some of the best human brains on the planet and make progress on issues related to the Friendly AI problem, the problem with the greatest leverage in the universe. I have suggested that Less Wrong would make faster progress toward this goal if it worked more directly with the community of scholars already tackling the exact same problems. I don't personally work toward this goal because I'm not mathematically sophisticated enough to do so, but I'm glad others are!
Still, I think the first goal could be more explicitly pursued. There are many people like myself and jwhendy who can be massively impacted for the better not by coming to a realization about algorithmic learning theory, but by coming to understand the basics of rationality like probability and the proper role of belief and reductionism.
Reasons for Less Wrong to devote more energy to the basics
How to do it
Let me put some meat on this. What does more focus on the basics look like? Here are some ideas: