With a population of seven billion, it's hard not to commit a weak version of the typical mind fallacy and assume you can assume anything about people's motivations.
Diverse as we all are, humans occupy but a tiny pin-point in the mind-design space of all possible intelligent agents. That our values are mostly aligned is one reason that there are so many opportunities for the positive-sum exchanges that have fueled the economic growth of the past several hundred years. I rarely encounter anyone that really truly desires a future that I find horrible (though plenty claim to want such futures). On the other hand, plenty of people go about achieving their goals in all sorts of awfully inefficient ways.
Improving the world's rationality to the point that people can actually achieve their goals seems like the least of our worries.
In which I worry that the Less Wrong project might go horribly right. This post belongs to my Altruist Support sequence.
Every project needs a risk assessment.
There's a feeling, just bubbling under the surface here at Less Wrong, that we're just playing at rationality. It's rationality kindergarten. The problem has been expressed in various ways:
And people are starting to look at fixing it. I'm not worried that their attempts - and mine - will fail. At least we'd have fun and learn something.
I'm worried that they will succeed.
What would such a Super Less Wrong community do? Its members would self-improve to the point where they had a good chance of succeeding at most things they put their mind to. They would recruit new rationalists and then optimize that recruitment process, until the community got big. They would develop methods for rapidly generating, classifying and evaluating ideas, so that the only ideas that got tried would be the best that anyone had come up with so far. The group would structure itself so that people's basic social drives - such as their desire for status - worked in the interests of the group rather than against it.
It would be pretty formidable.
What would the products of such a community be? There would probably be a self-help book that works. There would be an effective, practical guide to setting up effective communities. There would be an intuitive, practical guide to human behavior. There would be books, seminars and classes on how to really achieve your goals - and only the materials which actually got results would be kept. There would be a bunch of stuff on the Dark Arts too, no doubt. Possibly some AI research.
That's a whole lot of material that we wouldn't want to get into the hands of the wrong people.
Dangers include:
If this is a problem we should take seriously, what are some possible strategies for dealing with it?
In the post title, I have suggested an analogy with AI takeoff. That's not entirely fair; there is probably an upper bound to how effective a community of humans can be, at least until brain implants come along. We're probably talking two orders of magnitude rather than ten. But given that humanity already has technology with slight existential threat implications (nuclear weapons, rudimentary AI research), I would be worried about a movement that aims to make all of humanity more effective at everything they do.